mediapipe-nightly 0.10.21.post20241223__cp39-cp39-manylinux_2_28_x86_64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (593) hide show
  1. mediapipe/__init__.py +26 -0
  2. mediapipe/calculators/__init__.py +0 -0
  3. mediapipe/calculators/audio/__init__.py +0 -0
  4. mediapipe/calculators/audio/mfcc_mel_calculators_pb2.py +33 -0
  5. mediapipe/calculators/audio/rational_factor_resample_calculator_pb2.py +33 -0
  6. mediapipe/calculators/audio/spectrogram_calculator_pb2.py +37 -0
  7. mediapipe/calculators/audio/stabilized_log_calculator_pb2.py +31 -0
  8. mediapipe/calculators/audio/time_series_framer_calculator_pb2.py +33 -0
  9. mediapipe/calculators/core/__init__.py +0 -0
  10. mediapipe/calculators/core/bypass_calculator_pb2.py +31 -0
  11. mediapipe/calculators/core/clip_vector_size_calculator_pb2.py +31 -0
  12. mediapipe/calculators/core/concatenate_vector_calculator_pb2.py +31 -0
  13. mediapipe/calculators/core/constant_side_packet_calculator_pb2.py +39 -0
  14. mediapipe/calculators/core/dequantize_byte_array_calculator_pb2.py +31 -0
  15. mediapipe/calculators/core/flow_limiter_calculator_pb2.py +32 -0
  16. mediapipe/calculators/core/gate_calculator_pb2.py +33 -0
  17. mediapipe/calculators/core/get_vector_item_calculator_pb2.py +31 -0
  18. mediapipe/calculators/core/graph_profile_calculator_pb2.py +31 -0
  19. mediapipe/calculators/core/packet_cloner_calculator_pb2.py +31 -0
  20. mediapipe/calculators/core/packet_resampler_calculator_pb2.py +33 -0
  21. mediapipe/calculators/core/packet_thinner_calculator_pb2.py +33 -0
  22. mediapipe/calculators/core/quantize_float_vector_calculator_pb2.py +31 -0
  23. mediapipe/calculators/core/sequence_shift_calculator_pb2.py +31 -0
  24. mediapipe/calculators/core/split_vector_calculator_pb2.py +33 -0
  25. mediapipe/calculators/image/__init__.py +0 -0
  26. mediapipe/calculators/image/bilateral_filter_calculator_pb2.py +31 -0
  27. mediapipe/calculators/image/feature_detector_calculator_pb2.py +31 -0
  28. mediapipe/calculators/image/image_clone_calculator_pb2.py +31 -0
  29. mediapipe/calculators/image/image_cropping_calculator_pb2.py +33 -0
  30. mediapipe/calculators/image/image_transformation_calculator_pb2.py +38 -0
  31. mediapipe/calculators/image/mask_overlay_calculator_pb2.py +33 -0
  32. mediapipe/calculators/image/opencv_encoded_image_to_image_frame_calculator_pb2.py +31 -0
  33. mediapipe/calculators/image/opencv_image_encoder_calculator_pb2.py +35 -0
  34. mediapipe/calculators/image/recolor_calculator_pb2.py +34 -0
  35. mediapipe/calculators/image/rotation_mode_pb2.py +29 -0
  36. mediapipe/calculators/image/scale_image_calculator_pb2.py +34 -0
  37. mediapipe/calculators/image/segmentation_smoothing_calculator_pb2.py +31 -0
  38. mediapipe/calculators/image/set_alpha_calculator_pb2.py +31 -0
  39. mediapipe/calculators/image/warp_affine_calculator_pb2.py +36 -0
  40. mediapipe/calculators/internal/__init__.py +0 -0
  41. mediapipe/calculators/internal/callback_packet_calculator_pb2.py +33 -0
  42. mediapipe/calculators/tensor/__init__.py +0 -0
  43. mediapipe/calculators/tensor/audio_to_tensor_calculator_pb2.py +35 -0
  44. mediapipe/calculators/tensor/bert_preprocessor_calculator_pb2.py +31 -0
  45. mediapipe/calculators/tensor/feedback_tensors_calculator_pb2.py +37 -0
  46. mediapipe/calculators/tensor/image_to_tensor_calculator_pb2.py +40 -0
  47. mediapipe/calculators/tensor/inference_calculator_pb2.py +63 -0
  48. mediapipe/calculators/tensor/landmarks_to_tensor_calculator_pb2.py +33 -0
  49. mediapipe/calculators/tensor/regex_preprocessor_calculator_pb2.py +31 -0
  50. mediapipe/calculators/tensor/tensor_converter_calculator_pb2.py +34 -0
  51. mediapipe/calculators/tensor/tensor_to_joints_calculator_pb2.py +31 -0
  52. mediapipe/calculators/tensor/tensors_readback_calculator_pb2.py +35 -0
  53. mediapipe/calculators/tensor/tensors_to_audio_calculator_pb2.py +33 -0
  54. mediapipe/calculators/tensor/tensors_to_classification_calculator_pb2.py +44 -0
  55. mediapipe/calculators/tensor/tensors_to_detections_calculator_pb2.py +39 -0
  56. mediapipe/calculators/tensor/tensors_to_floats_calculator_pb2.py +33 -0
  57. mediapipe/calculators/tensor/tensors_to_landmarks_calculator_pb2.py +33 -0
  58. mediapipe/calculators/tensor/tensors_to_segmentation_calculator_pb2.py +34 -0
  59. mediapipe/calculators/tensor/vector_to_tensor_calculator_pb2.py +27 -0
  60. mediapipe/calculators/tflite/__init__.py +0 -0
  61. mediapipe/calculators/tflite/ssd_anchors_calculator_pb2.py +32 -0
  62. mediapipe/calculators/tflite/tflite_converter_calculator_pb2.py +33 -0
  63. mediapipe/calculators/tflite/tflite_custom_op_resolver_calculator_pb2.py +31 -0
  64. mediapipe/calculators/tflite/tflite_inference_calculator_pb2.py +49 -0
  65. mediapipe/calculators/tflite/tflite_tensors_to_classification_calculator_pb2.py +31 -0
  66. mediapipe/calculators/tflite/tflite_tensors_to_detections_calculator_pb2.py +31 -0
  67. mediapipe/calculators/tflite/tflite_tensors_to_landmarks_calculator_pb2.py +33 -0
  68. mediapipe/calculators/tflite/tflite_tensors_to_segmentation_calculator_pb2.py +31 -0
  69. mediapipe/calculators/util/__init__.py +0 -0
  70. mediapipe/calculators/util/align_hand_to_pose_in_world_calculator_pb2.py +31 -0
  71. mediapipe/calculators/util/annotation_overlay_calculator_pb2.py +32 -0
  72. mediapipe/calculators/util/association_calculator_pb2.py +31 -0
  73. mediapipe/calculators/util/collection_has_min_size_calculator_pb2.py +31 -0
  74. mediapipe/calculators/util/combine_joints_calculator_pb2.py +36 -0
  75. mediapipe/calculators/util/detection_label_id_to_text_calculator_pb2.py +36 -0
  76. mediapipe/calculators/util/detections_to_rects_calculator_pb2.py +33 -0
  77. mediapipe/calculators/util/detections_to_render_data_calculator_pb2.py +33 -0
  78. mediapipe/calculators/util/face_to_rect_calculator_pb2.py +26 -0
  79. mediapipe/calculators/util/filter_detections_calculator_pb2.py +31 -0
  80. mediapipe/calculators/util/flat_color_image_calculator_pb2.py +32 -0
  81. mediapipe/calculators/util/labels_to_render_data_calculator_pb2.py +34 -0
  82. mediapipe/calculators/util/landmark_projection_calculator_pb2.py +31 -0
  83. mediapipe/calculators/util/landmarks_refinement_calculator_pb2.py +41 -0
  84. mediapipe/calculators/util/landmarks_smoothing_calculator_pb2.py +33 -0
  85. mediapipe/calculators/util/landmarks_to_detection_calculator_pb2.py +31 -0
  86. mediapipe/calculators/util/landmarks_to_floats_calculator_pb2.py +31 -0
  87. mediapipe/calculators/util/landmarks_to_render_data_calculator_pb2.py +32 -0
  88. mediapipe/calculators/util/landmarks_transformation_calculator_pb2.py +37 -0
  89. mediapipe/calculators/util/latency_pb2.py +26 -0
  90. mediapipe/calculators/util/local_file_contents_calculator_pb2.py +31 -0
  91. mediapipe/calculators/util/logic_calculator_pb2.py +34 -0
  92. mediapipe/calculators/util/non_max_suppression_calculator_pb2.py +35 -0
  93. mediapipe/calculators/util/packet_frequency_calculator_pb2.py +31 -0
  94. mediapipe/calculators/util/packet_frequency_pb2.py +26 -0
  95. mediapipe/calculators/util/packet_latency_calculator_pb2.py +31 -0
  96. mediapipe/calculators/util/rect_to_render_data_calculator_pb2.py +32 -0
  97. mediapipe/calculators/util/rect_to_render_scale_calculator_pb2.py +31 -0
  98. mediapipe/calculators/util/rect_transformation_calculator_pb2.py +31 -0
  99. mediapipe/calculators/util/refine_landmarks_from_heatmap_calculator_pb2.py +31 -0
  100. mediapipe/calculators/util/resource_provider_calculator_pb2.py +28 -0
  101. mediapipe/calculators/util/set_joints_visibility_calculator_pb2.py +41 -0
  102. mediapipe/calculators/util/thresholding_calculator_pb2.py +31 -0
  103. mediapipe/calculators/util/timed_box_list_id_to_label_calculator_pb2.py +31 -0
  104. mediapipe/calculators/util/timed_box_list_to_render_data_calculator_pb2.py +32 -0
  105. mediapipe/calculators/util/top_k_scores_calculator_pb2.py +31 -0
  106. mediapipe/calculators/util/visibility_copy_calculator_pb2.py +27 -0
  107. mediapipe/calculators/util/visibility_smoothing_calculator_pb2.py +31 -0
  108. mediapipe/calculators/video/__init__.py +0 -0
  109. mediapipe/calculators/video/box_detector_calculator_pb2.py +32 -0
  110. mediapipe/calculators/video/box_tracker_calculator_pb2.py +32 -0
  111. mediapipe/calculators/video/flow_packager_calculator_pb2.py +32 -0
  112. mediapipe/calculators/video/flow_to_image_calculator_pb2.py +31 -0
  113. mediapipe/calculators/video/motion_analysis_calculator_pb2.py +42 -0
  114. mediapipe/calculators/video/opencv_video_encoder_calculator_pb2.py +31 -0
  115. mediapipe/calculators/video/tool/__init__.py +0 -0
  116. mediapipe/calculators/video/tool/flow_quantizer_model_pb2.py +26 -0
  117. mediapipe/calculators/video/tracked_detection_manager_calculator_pb2.py +32 -0
  118. mediapipe/calculators/video/video_pre_stream_calculator_pb2.py +35 -0
  119. mediapipe/examples/__init__.py +14 -0
  120. mediapipe/examples/desktop/__init__.py +14 -0
  121. mediapipe/framework/__init__.py +0 -0
  122. mediapipe/framework/calculator_options_pb2.py +29 -0
  123. mediapipe/framework/calculator_pb2.py +59 -0
  124. mediapipe/framework/calculator_profile_pb2.py +48 -0
  125. mediapipe/framework/deps/__init__.py +0 -0
  126. mediapipe/framework/deps/proto_descriptor_pb2.py +29 -0
  127. mediapipe/framework/formats/__init__.py +0 -0
  128. mediapipe/framework/formats/affine_transform_data_pb2.py +28 -0
  129. mediapipe/framework/formats/annotation/__init__.py +0 -0
  130. mediapipe/framework/formats/annotation/locus_pb2.py +32 -0
  131. mediapipe/framework/formats/annotation/rasterization_pb2.py +29 -0
  132. mediapipe/framework/formats/body_rig_pb2.py +28 -0
  133. mediapipe/framework/formats/classification_pb2.py +31 -0
  134. mediapipe/framework/formats/detection_pb2.py +36 -0
  135. mediapipe/framework/formats/image_file_properties_pb2.py +26 -0
  136. mediapipe/framework/formats/image_format_pb2.py +29 -0
  137. mediapipe/framework/formats/landmark_pb2.py +37 -0
  138. mediapipe/framework/formats/location_data_pb2.py +38 -0
  139. mediapipe/framework/formats/matrix_data_pb2.py +31 -0
  140. mediapipe/framework/formats/motion/__init__.py +0 -0
  141. mediapipe/framework/formats/motion/optical_flow_field_data_pb2.py +30 -0
  142. mediapipe/framework/formats/object_detection/__init__.py +0 -0
  143. mediapipe/framework/formats/object_detection/anchor_pb2.py +26 -0
  144. mediapipe/framework/formats/rect_pb2.py +29 -0
  145. mediapipe/framework/formats/time_series_header_pb2.py +28 -0
  146. mediapipe/framework/graph_runtime_info_pb2.py +31 -0
  147. mediapipe/framework/mediapipe_options_pb2.py +27 -0
  148. mediapipe/framework/packet_factory_pb2.py +31 -0
  149. mediapipe/framework/packet_generator_pb2.py +33 -0
  150. mediapipe/framework/status_handler_pb2.py +28 -0
  151. mediapipe/framework/stream_handler/__init__.py +0 -0
  152. mediapipe/framework/stream_handler/default_input_stream_handler_pb2.py +27 -0
  153. mediapipe/framework/stream_handler/fixed_size_input_stream_handler_pb2.py +27 -0
  154. mediapipe/framework/stream_handler/sync_set_input_stream_handler_pb2.py +29 -0
  155. mediapipe/framework/stream_handler/timestamp_align_input_stream_handler_pb2.py +27 -0
  156. mediapipe/framework/stream_handler_pb2.py +30 -0
  157. mediapipe/framework/test_calculators_pb2.py +31 -0
  158. mediapipe/framework/thread_pool_executor_pb2.py +29 -0
  159. mediapipe/framework/tool/__init__.py +0 -0
  160. mediapipe/framework/tool/calculator_graph_template_pb2.py +44 -0
  161. mediapipe/framework/tool/field_data_pb2.py +28 -0
  162. mediapipe/framework/tool/node_chain_subgraph_pb2.py +31 -0
  163. mediapipe/framework/tool/packet_generator_wrapper_calculator_pb2.py +28 -0
  164. mediapipe/framework/tool/source_pb2.py +33 -0
  165. mediapipe/framework/tool/switch_container_pb2.py +32 -0
  166. mediapipe/gpu/__init__.py +0 -0
  167. mediapipe/gpu/copy_calculator_pb2.py +33 -0
  168. mediapipe/gpu/gl_animation_overlay_calculator_pb2.py +31 -0
  169. mediapipe/gpu/gl_context_options_pb2.py +31 -0
  170. mediapipe/gpu/gl_scaler_calculator_pb2.py +32 -0
  171. mediapipe/gpu/gl_surface_sink_calculator_pb2.py +32 -0
  172. mediapipe/gpu/gpu_origin_pb2.py +29 -0
  173. mediapipe/gpu/scale_mode_pb2.py +28 -0
  174. mediapipe/model_maker/__init__.py +27 -0
  175. mediapipe/model_maker/setup.py +107 -0
  176. mediapipe/modules/__init__.py +0 -0
  177. mediapipe/modules/face_detection/__init__.py +0 -0
  178. mediapipe/modules/face_detection/face_detection_full_range_cpu.binarypb +0 -0
  179. mediapipe/modules/face_detection/face_detection_full_range_sparse.tflite +0 -0
  180. mediapipe/modules/face_detection/face_detection_pb2.py +30 -0
  181. mediapipe/modules/face_detection/face_detection_short_range.tflite +0 -0
  182. mediapipe/modules/face_detection/face_detection_short_range_cpu.binarypb +0 -0
  183. mediapipe/modules/face_geometry/__init__.py +0 -0
  184. mediapipe/modules/face_geometry/data/__init__.py +0 -0
  185. mediapipe/modules/face_geometry/effect_renderer_calculator_pb2.py +27 -0
  186. mediapipe/modules/face_geometry/env_generator_calculator_pb2.py +28 -0
  187. mediapipe/modules/face_geometry/geometry_pipeline_calculator_pb2.py +27 -0
  188. mediapipe/modules/face_geometry/libs/__init__.py +0 -0
  189. mediapipe/modules/face_geometry/protos/__init__.py +0 -0
  190. mediapipe/modules/face_geometry/protos/environment_pb2.py +31 -0
  191. mediapipe/modules/face_geometry/protos/face_geometry_pb2.py +29 -0
  192. mediapipe/modules/face_geometry/protos/geometry_pipeline_metadata_pb2.py +32 -0
  193. mediapipe/modules/face_geometry/protos/mesh_3d_pb2.py +31 -0
  194. mediapipe/modules/face_landmark/__init__.py +0 -0
  195. mediapipe/modules/face_landmark/face_landmark.tflite +0 -0
  196. mediapipe/modules/face_landmark/face_landmark_front_cpu.binarypb +0 -0
  197. mediapipe/modules/face_landmark/face_landmark_with_attention.tflite +0 -0
  198. mediapipe/modules/hand_landmark/__init__.py +0 -0
  199. mediapipe/modules/hand_landmark/calculators/__init__.py +0 -0
  200. mediapipe/modules/hand_landmark/hand_landmark_full.tflite +0 -0
  201. mediapipe/modules/hand_landmark/hand_landmark_lite.tflite +0 -0
  202. mediapipe/modules/hand_landmark/hand_landmark_tracking_cpu.binarypb +0 -0
  203. mediapipe/modules/hand_landmark/handedness.txt +2 -0
  204. mediapipe/modules/holistic_landmark/__init__.py +0 -0
  205. mediapipe/modules/holistic_landmark/calculators/__init__.py +0 -0
  206. mediapipe/modules/holistic_landmark/calculators/roi_tracking_calculator_pb2.py +37 -0
  207. mediapipe/modules/holistic_landmark/hand_recrop.tflite +0 -0
  208. mediapipe/modules/holistic_landmark/holistic_landmark_cpu.binarypb +0 -0
  209. mediapipe/modules/iris_landmark/__init__.py +0 -0
  210. mediapipe/modules/iris_landmark/iris_landmark.tflite +0 -0
  211. mediapipe/modules/objectron/__init__.py +0 -0
  212. mediapipe/modules/objectron/calculators/__init__.py +0 -0
  213. mediapipe/modules/objectron/calculators/a_r_capture_metadata_pb2.py +102 -0
  214. mediapipe/modules/objectron/calculators/annotation_data_pb2.py +38 -0
  215. mediapipe/modules/objectron/calculators/belief_decoder_config_pb2.py +28 -0
  216. mediapipe/modules/objectron/calculators/camera_parameters_pb2.py +30 -0
  217. mediapipe/modules/objectron/calculators/filter_detection_calculator_pb2.py +35 -0
  218. mediapipe/modules/objectron/calculators/frame_annotation_to_rect_calculator_pb2.py +31 -0
  219. mediapipe/modules/objectron/calculators/frame_annotation_tracker_calculator_pb2.py +31 -0
  220. mediapipe/modules/objectron/calculators/lift_2d_frame_annotation_to_3d_calculator_pb2.py +32 -0
  221. mediapipe/modules/objectron/calculators/object_pb2.py +38 -0
  222. mediapipe/modules/objectron/calculators/tensors_to_objects_calculator_pb2.py +32 -0
  223. mediapipe/modules/objectron/calculators/tflite_tensors_to_objects_calculator_pb2.py +32 -0
  224. mediapipe/modules/objectron/object_detection_oidv4_labelmap.txt +24 -0
  225. mediapipe/modules/objectron/objectron_cpu.binarypb +0 -0
  226. mediapipe/modules/palm_detection/__init__.py +0 -0
  227. mediapipe/modules/palm_detection/palm_detection_full.tflite +0 -0
  228. mediapipe/modules/palm_detection/palm_detection_lite.tflite +0 -0
  229. mediapipe/modules/pose_detection/__init__.py +0 -0
  230. mediapipe/modules/pose_detection/pose_detection.tflite +0 -0
  231. mediapipe/modules/pose_landmark/__init__.py +0 -0
  232. mediapipe/modules/pose_landmark/pose_landmark_cpu.binarypb +0 -0
  233. mediapipe/modules/pose_landmark/pose_landmark_full.tflite +0 -0
  234. mediapipe/modules/selfie_segmentation/__init__.py +0 -0
  235. mediapipe/modules/selfie_segmentation/selfie_segmentation.tflite +0 -0
  236. mediapipe/modules/selfie_segmentation/selfie_segmentation_cpu.binarypb +0 -0
  237. mediapipe/modules/selfie_segmentation/selfie_segmentation_landscape.tflite +0 -0
  238. mediapipe/python/__init__.py +29 -0
  239. mediapipe/python/_framework_bindings.cpython-39-x86_64-linux-gnu.so +0 -0
  240. mediapipe/python/calculator_graph_test.py +251 -0
  241. mediapipe/python/image_frame_test.py +194 -0
  242. mediapipe/python/image_test.py +218 -0
  243. mediapipe/python/packet_creator.py +275 -0
  244. mediapipe/python/packet_getter.py +120 -0
  245. mediapipe/python/packet_test.py +533 -0
  246. mediapipe/python/solution_base.py +604 -0
  247. mediapipe/python/solution_base_test.py +396 -0
  248. mediapipe/python/solutions/__init__.py +27 -0
  249. mediapipe/python/solutions/download_utils.py +37 -0
  250. mediapipe/python/solutions/drawing_styles.py +249 -0
  251. mediapipe/python/solutions/drawing_utils.py +320 -0
  252. mediapipe/python/solutions/drawing_utils_test.py +258 -0
  253. mediapipe/python/solutions/face_detection.py +105 -0
  254. mediapipe/python/solutions/face_detection_test.py +92 -0
  255. mediapipe/python/solutions/face_mesh.py +125 -0
  256. mediapipe/python/solutions/face_mesh_connections.py +500 -0
  257. mediapipe/python/solutions/face_mesh_test.py +170 -0
  258. mediapipe/python/solutions/hands.py +153 -0
  259. mediapipe/python/solutions/hands_connections.py +32 -0
  260. mediapipe/python/solutions/hands_test.py +219 -0
  261. mediapipe/python/solutions/holistic.py +167 -0
  262. mediapipe/python/solutions/holistic_test.py +142 -0
  263. mediapipe/python/solutions/objectron.py +288 -0
  264. mediapipe/python/solutions/objectron_test.py +81 -0
  265. mediapipe/python/solutions/pose.py +192 -0
  266. mediapipe/python/solutions/pose_connections.py +22 -0
  267. mediapipe/python/solutions/pose_test.py +262 -0
  268. mediapipe/python/solutions/selfie_segmentation.py +76 -0
  269. mediapipe/python/solutions/selfie_segmentation_test.py +68 -0
  270. mediapipe/python/timestamp_test.py +78 -0
  271. mediapipe/tasks/__init__.py +14 -0
  272. mediapipe/tasks/cc/__init__.py +0 -0
  273. mediapipe/tasks/cc/audio/__init__.py +0 -0
  274. mediapipe/tasks/cc/audio/audio_classifier/__init__.py +0 -0
  275. mediapipe/tasks/cc/audio/audio_classifier/proto/__init__.py +0 -0
  276. mediapipe/tasks/cc/audio/audio_classifier/proto/audio_classifier_graph_options_pb2.py +35 -0
  277. mediapipe/tasks/cc/audio/audio_embedder/__init__.py +0 -0
  278. mediapipe/tasks/cc/audio/audio_embedder/proto/__init__.py +0 -0
  279. mediapipe/tasks/cc/audio/audio_embedder/proto/audio_embedder_graph_options_pb2.py +35 -0
  280. mediapipe/tasks/cc/audio/core/__init__.py +0 -0
  281. mediapipe/tasks/cc/audio/utils/__init__.py +0 -0
  282. mediapipe/tasks/cc/components/__init__.py +0 -0
  283. mediapipe/tasks/cc/components/calculators/__init__.py +0 -0
  284. mediapipe/tasks/cc/components/calculators/classification_aggregation_calculator_pb2.py +31 -0
  285. mediapipe/tasks/cc/components/calculators/score_calibration_calculator_pb2.py +35 -0
  286. mediapipe/tasks/cc/components/calculators/tensors_to_embeddings_calculator_pb2.py +32 -0
  287. mediapipe/tasks/cc/components/containers/__init__.py +0 -0
  288. mediapipe/tasks/cc/components/containers/proto/__init__.py +0 -0
  289. mediapipe/tasks/cc/components/containers/proto/classifications_pb2.py +30 -0
  290. mediapipe/tasks/cc/components/containers/proto/embeddings_pb2.py +35 -0
  291. mediapipe/tasks/cc/components/containers/proto/landmarks_detection_result_pb2.py +32 -0
  292. mediapipe/tasks/cc/components/processors/__init__.py +0 -0
  293. mediapipe/tasks/cc/components/processors/proto/__init__.py +0 -0
  294. mediapipe/tasks/cc/components/processors/proto/classification_postprocessing_graph_options_pb2.py +38 -0
  295. mediapipe/tasks/cc/components/processors/proto/classifier_options_pb2.py +27 -0
  296. mediapipe/tasks/cc/components/processors/proto/detection_postprocessing_graph_options_pb2.py +36 -0
  297. mediapipe/tasks/cc/components/processors/proto/detector_options_pb2.py +27 -0
  298. mediapipe/tasks/cc/components/processors/proto/embedder_options_pb2.py +27 -0
  299. mediapipe/tasks/cc/components/processors/proto/embedding_postprocessing_graph_options_pb2.py +32 -0
  300. mediapipe/tasks/cc/components/processors/proto/image_preprocessing_graph_options_pb2.py +34 -0
  301. mediapipe/tasks/cc/components/processors/proto/text_model_type_pb2.py +28 -0
  302. mediapipe/tasks/cc/components/processors/proto/text_preprocessing_graph_options_pb2.py +32 -0
  303. mediapipe/tasks/cc/components/utils/__init__.py +0 -0
  304. mediapipe/tasks/cc/core/__init__.py +0 -0
  305. mediapipe/tasks/cc/core/proto/__init__.py +0 -0
  306. mediapipe/tasks/cc/core/proto/acceleration_pb2.py +28 -0
  307. mediapipe/tasks/cc/core/proto/base_options_pb2.py +30 -0
  308. mediapipe/tasks/cc/core/proto/external_file_pb2.py +31 -0
  309. mediapipe/tasks/cc/core/proto/inference_subgraph_pb2.py +32 -0
  310. mediapipe/tasks/cc/core/proto/model_resources_calculator_pb2.py +32 -0
  311. mediapipe/tasks/cc/genai/__init__.py +0 -0
  312. mediapipe/tasks/cc/genai/inference/__init__.py +0 -0
  313. mediapipe/tasks/cc/genai/inference/c/__init__.py +0 -0
  314. mediapipe/tasks/cc/genai/inference/calculators/__init__.py +0 -0
  315. mediapipe/tasks/cc/genai/inference/calculators/detokenizer_calculator_pb2.py +27 -0
  316. mediapipe/tasks/cc/genai/inference/calculators/llm_gpu_calculator_pb2.py +32 -0
  317. mediapipe/tasks/cc/genai/inference/calculators/model_data_calculator_pb2.py +27 -0
  318. mediapipe/tasks/cc/genai/inference/calculators/tokenizer_calculator_pb2.py +29 -0
  319. mediapipe/tasks/cc/genai/inference/common/__init__.py +0 -0
  320. mediapipe/tasks/cc/genai/inference/proto/__init__.py +0 -0
  321. mediapipe/tasks/cc/genai/inference/proto/llm_file_metadata_pb2.py +32 -0
  322. mediapipe/tasks/cc/genai/inference/proto/llm_params_pb2.py +33 -0
  323. mediapipe/tasks/cc/genai/inference/proto/prompt_template_pb2.py +27 -0
  324. mediapipe/tasks/cc/genai/inference/proto/sampler_params_pb2.py +29 -0
  325. mediapipe/tasks/cc/genai/inference/proto/transformer_params_pb2.py +45 -0
  326. mediapipe/tasks/cc/genai/inference/utils/__init__.py +0 -0
  327. mediapipe/tasks/cc/genai/inference/utils/llm_utils/__init__.py +0 -0
  328. mediapipe/tasks/cc/genai/inference/utils/xnn_utils/__init__.py +0 -0
  329. mediapipe/tasks/cc/metadata/__init__.py +0 -0
  330. mediapipe/tasks/cc/metadata/python/__init__.py +0 -0
  331. mediapipe/tasks/cc/metadata/python/_pywrap_metadata_version.cpython-39-x86_64-linux-gnu.so +0 -0
  332. mediapipe/tasks/cc/metadata/tests/__init__.py +0 -0
  333. mediapipe/tasks/cc/metadata/utils/__init__.py +0 -0
  334. mediapipe/tasks/cc/text/__init__.py +0 -0
  335. mediapipe/tasks/cc/text/custom_ops/__init__.py +0 -0
  336. mediapipe/tasks/cc/text/custom_ops/ragged/__init__.py +0 -0
  337. mediapipe/tasks/cc/text/custom_ops/sentencepiece/__init__.py +0 -0
  338. mediapipe/tasks/cc/text/custom_ops/sentencepiece/testdata/__init__.py +0 -0
  339. mediapipe/tasks/cc/text/language_detector/__init__.py +0 -0
  340. mediapipe/tasks/cc/text/language_detector/custom_ops/__init__.py +0 -0
  341. mediapipe/tasks/cc/text/language_detector/custom_ops/utils/__init__.py +0 -0
  342. mediapipe/tasks/cc/text/language_detector/custom_ops/utils/hash/__init__.py +0 -0
  343. mediapipe/tasks/cc/text/language_detector/custom_ops/utils/utf/__init__.py +0 -0
  344. mediapipe/tasks/cc/text/text_classifier/__init__.py +0 -0
  345. mediapipe/tasks/cc/text/text_classifier/proto/__init__.py +0 -0
  346. mediapipe/tasks/cc/text/text_classifier/proto/text_classifier_graph_options_pb2.py +35 -0
  347. mediapipe/tasks/cc/text/text_embedder/__init__.py +0 -0
  348. mediapipe/tasks/cc/text/text_embedder/proto/__init__.py +0 -0
  349. mediapipe/tasks/cc/text/text_embedder/proto/text_embedder_graph_options_pb2.py +35 -0
  350. mediapipe/tasks/cc/text/tokenizers/__init__.py +0 -0
  351. mediapipe/tasks/cc/text/utils/__init__.py +0 -0
  352. mediapipe/tasks/cc/vision/__init__.py +0 -0
  353. mediapipe/tasks/cc/vision/core/__init__.py +0 -0
  354. mediapipe/tasks/cc/vision/custom_ops/__init__.py +0 -0
  355. mediapipe/tasks/cc/vision/face_detector/__init__.py +0 -0
  356. mediapipe/tasks/cc/vision/face_detector/proto/__init__.py +0 -0
  357. mediapipe/tasks/cc/vision/face_detector/proto/face_detector_graph_options_pb2.py +34 -0
  358. mediapipe/tasks/cc/vision/face_geometry/__init__.py +0 -0
  359. mediapipe/tasks/cc/vision/face_geometry/calculators/__init__.py +0 -0
  360. mediapipe/tasks/cc/vision/face_geometry/calculators/env_generator_calculator_pb2.py +28 -0
  361. mediapipe/tasks/cc/vision/face_geometry/calculators/geometry_pipeline_calculator_pb2.py +29 -0
  362. mediapipe/tasks/cc/vision/face_geometry/data/__init__.py +0 -0
  363. mediapipe/tasks/cc/vision/face_geometry/libs/__init__.py +0 -0
  364. mediapipe/tasks/cc/vision/face_geometry/proto/__init__.py +0 -0
  365. mediapipe/tasks/cc/vision/face_geometry/proto/environment_pb2.py +31 -0
  366. mediapipe/tasks/cc/vision/face_geometry/proto/face_geometry_graph_options_pb2.py +29 -0
  367. mediapipe/tasks/cc/vision/face_geometry/proto/face_geometry_pb2.py +29 -0
  368. mediapipe/tasks/cc/vision/face_geometry/proto/geometry_pipeline_metadata_pb2.py +32 -0
  369. mediapipe/tasks/cc/vision/face_geometry/proto/mesh_3d_pb2.py +31 -0
  370. mediapipe/tasks/cc/vision/face_landmarker/__init__.py +0 -0
  371. mediapipe/tasks/cc/vision/face_landmarker/proto/__init__.py +0 -0
  372. mediapipe/tasks/cc/vision/face_landmarker/proto/face_blendshapes_graph_options_pb2.py +34 -0
  373. mediapipe/tasks/cc/vision/face_landmarker/proto/face_landmarker_graph_options_pb2.py +37 -0
  374. mediapipe/tasks/cc/vision/face_landmarker/proto/face_landmarks_detector_graph_options_pb2.py +35 -0
  375. mediapipe/tasks/cc/vision/face_landmarker/proto/tensors_to_face_landmarks_graph_options_pb2.py +32 -0
  376. mediapipe/tasks/cc/vision/face_stylizer/__init__.py +0 -0
  377. mediapipe/tasks/cc/vision/face_stylizer/calculators/__init__.py +0 -0
  378. mediapipe/tasks/cc/vision/face_stylizer/calculators/tensors_to_image_calculator_pb2.py +36 -0
  379. mediapipe/tasks/cc/vision/face_stylizer/proto/__init__.py +0 -0
  380. mediapipe/tasks/cc/vision/face_stylizer/proto/face_stylizer_graph_options_pb2.py +35 -0
  381. mediapipe/tasks/cc/vision/gesture_recognizer/__init__.py +0 -0
  382. mediapipe/tasks/cc/vision/gesture_recognizer/calculators/__init__.py +0 -0
  383. mediapipe/tasks/cc/vision/gesture_recognizer/calculators/combined_prediction_calculator_pb2.py +33 -0
  384. mediapipe/tasks/cc/vision/gesture_recognizer/calculators/landmarks_to_matrix_calculator_pb2.py +31 -0
  385. mediapipe/tasks/cc/vision/gesture_recognizer/proto/__init__.py +0 -0
  386. mediapipe/tasks/cc/vision/gesture_recognizer/proto/gesture_classifier_graph_options_pb2.py +35 -0
  387. mediapipe/tasks/cc/vision/gesture_recognizer/proto/gesture_embedder_graph_options_pb2.py +34 -0
  388. mediapipe/tasks/cc/vision/gesture_recognizer/proto/gesture_recognizer_graph_options_pb2.py +36 -0
  389. mediapipe/tasks/cc/vision/gesture_recognizer/proto/hand_gesture_recognizer_graph_options_pb2.py +36 -0
  390. mediapipe/tasks/cc/vision/hand_detector/__init__.py +0 -0
  391. mediapipe/tasks/cc/vision/hand_detector/proto/__init__.py +0 -0
  392. mediapipe/tasks/cc/vision/hand_detector/proto/hand_detector_graph_options_pb2.py +34 -0
  393. mediapipe/tasks/cc/vision/hand_detector/proto/hand_detector_result_pb2.py +30 -0
  394. mediapipe/tasks/cc/vision/hand_landmarker/__init__.py +0 -0
  395. mediapipe/tasks/cc/vision/hand_landmarker/calculators/__init__.py +0 -0
  396. mediapipe/tasks/cc/vision/hand_landmarker/calculators/hand_association_calculator_pb2.py +31 -0
  397. mediapipe/tasks/cc/vision/hand_landmarker/proto/__init__.py +0 -0
  398. mediapipe/tasks/cc/vision/hand_landmarker/proto/hand_landmarker_graph_options_pb2.py +36 -0
  399. mediapipe/tasks/cc/vision/hand_landmarker/proto/hand_landmarks_detector_graph_options_pb2.py +34 -0
  400. mediapipe/tasks/cc/vision/hand_landmarker/proto/hand_roi_refinement_graph_options_pb2.py +28 -0
  401. mediapipe/tasks/cc/vision/holistic_landmarker/__init__.py +0 -0
  402. mediapipe/tasks/cc/vision/holistic_landmarker/proto/__init__.py +0 -0
  403. mediapipe/tasks/cc/vision/holistic_landmarker/proto/holistic_landmarker_graph_options_pb2.py +34 -0
  404. mediapipe/tasks/cc/vision/holistic_landmarker/proto/holistic_result_pb2.py +29 -0
  405. mediapipe/tasks/cc/vision/image_classifier/__init__.py +0 -0
  406. mediapipe/tasks/cc/vision/image_classifier/proto/__init__.py +0 -0
  407. mediapipe/tasks/cc/vision/image_classifier/proto/image_classifier_graph_options_pb2.py +35 -0
  408. mediapipe/tasks/cc/vision/image_embedder/__init__.py +0 -0
  409. mediapipe/tasks/cc/vision/image_embedder/proto/__init__.py +0 -0
  410. mediapipe/tasks/cc/vision/image_embedder/proto/image_embedder_graph_options_pb2.py +35 -0
  411. mediapipe/tasks/cc/vision/image_generator/__init__.py +0 -0
  412. mediapipe/tasks/cc/vision/image_generator/diffuser/__init__.py +0 -0
  413. mediapipe/tasks/cc/vision/image_generator/diffuser/stable_diffusion_iterate_calculator_pb2.py +40 -0
  414. mediapipe/tasks/cc/vision/image_generator/proto/__init__.py +0 -0
  415. mediapipe/tasks/cc/vision/image_generator/proto/conditioned_image_graph_options_pb2.py +40 -0
  416. mediapipe/tasks/cc/vision/image_generator/proto/control_plugin_graph_options_pb2.py +34 -0
  417. mediapipe/tasks/cc/vision/image_generator/proto/image_generator_graph_options_pb2.py +30 -0
  418. mediapipe/tasks/cc/vision/image_segmenter/__init__.py +0 -0
  419. mediapipe/tasks/cc/vision/image_segmenter/calculators/__init__.py +0 -0
  420. mediapipe/tasks/cc/vision/image_segmenter/calculators/tensors_to_segmentation_calculator_pb2.py +34 -0
  421. mediapipe/tasks/cc/vision/image_segmenter/proto/__init__.py +0 -0
  422. mediapipe/tasks/cc/vision/image_segmenter/proto/image_segmenter_graph_options_pb2.py +35 -0
  423. mediapipe/tasks/cc/vision/image_segmenter/proto/segmenter_options_pb2.py +33 -0
  424. mediapipe/tasks/cc/vision/interactive_segmenter/__init__.py +0 -0
  425. mediapipe/tasks/cc/vision/object_detector/__init__.py +0 -0
  426. mediapipe/tasks/cc/vision/object_detector/proto/__init__.py +0 -0
  427. mediapipe/tasks/cc/vision/object_detector/proto/object_detector_options_pb2.py +34 -0
  428. mediapipe/tasks/cc/vision/pose_detector/__init__.py +0 -0
  429. mediapipe/tasks/cc/vision/pose_detector/proto/__init__.py +0 -0
  430. mediapipe/tasks/cc/vision/pose_detector/proto/pose_detector_graph_options_pb2.py +34 -0
  431. mediapipe/tasks/cc/vision/pose_landmarker/__init__.py +0 -0
  432. mediapipe/tasks/cc/vision/pose_landmarker/proto/__init__.py +0 -0
  433. mediapipe/tasks/cc/vision/pose_landmarker/proto/pose_landmarker_graph_options_pb2.py +36 -0
  434. mediapipe/tasks/cc/vision/pose_landmarker/proto/pose_landmarks_detector_graph_options_pb2.py +34 -0
  435. mediapipe/tasks/cc/vision/utils/__init__.py +0 -0
  436. mediapipe/tasks/cc/vision/utils/ghum/__init__.py +0 -0
  437. mediapipe/tasks/metadata/image_segmenter_metadata_schema.fbs +59 -0
  438. mediapipe/tasks/metadata/image_segmenter_metadata_schema_py_generated.py +108 -0
  439. mediapipe/tasks/metadata/metadata_schema.fbs +732 -0
  440. mediapipe/tasks/metadata/metadata_schema_py_generated.py +3251 -0
  441. mediapipe/tasks/metadata/object_detector_metadata_schema.fbs +98 -0
  442. mediapipe/tasks/metadata/object_detector_metadata_schema_py_generated.py +674 -0
  443. mediapipe/tasks/metadata/schema_py_generated.py +18438 -0
  444. mediapipe/tasks/python/__init__.py +27 -0
  445. mediapipe/tasks/python/audio/__init__.py +33 -0
  446. mediapipe/tasks/python/audio/audio_classifier.py +324 -0
  447. mediapipe/tasks/python/audio/audio_embedder.py +285 -0
  448. mediapipe/tasks/python/audio/core/__init__.py +16 -0
  449. mediapipe/tasks/python/audio/core/audio_record.py +125 -0
  450. mediapipe/tasks/python/audio/core/audio_task_running_mode.py +29 -0
  451. mediapipe/tasks/python/audio/core/base_audio_task_api.py +181 -0
  452. mediapipe/tasks/python/benchmark/__init__.py +13 -0
  453. mediapipe/tasks/python/benchmark/benchmark_utils.py +70 -0
  454. mediapipe/tasks/python/benchmark/vision/__init__.py +13 -0
  455. mediapipe/tasks/python/benchmark/vision/benchmark.py +99 -0
  456. mediapipe/tasks/python/benchmark/vision/core/__init__.py +14 -0
  457. mediapipe/tasks/python/benchmark/vision/core/base_vision_benchmark_api.py +40 -0
  458. mediapipe/tasks/python/components/__init__.py +13 -0
  459. mediapipe/tasks/python/components/containers/__init__.py +53 -0
  460. mediapipe/tasks/python/components/containers/audio_data.py +137 -0
  461. mediapipe/tasks/python/components/containers/bounding_box.py +73 -0
  462. mediapipe/tasks/python/components/containers/category.py +78 -0
  463. mediapipe/tasks/python/components/containers/classification_result.py +111 -0
  464. mediapipe/tasks/python/components/containers/detections.py +181 -0
  465. mediapipe/tasks/python/components/containers/embedding_result.py +89 -0
  466. mediapipe/tasks/python/components/containers/keypoint.py +77 -0
  467. mediapipe/tasks/python/components/containers/landmark.py +122 -0
  468. mediapipe/tasks/python/components/containers/landmark_detection_result.py +106 -0
  469. mediapipe/tasks/python/components/containers/rect.py +109 -0
  470. mediapipe/tasks/python/components/processors/__init__.py +23 -0
  471. mediapipe/tasks/python/components/processors/classifier_options.py +86 -0
  472. mediapipe/tasks/python/components/utils/__init__.py +13 -0
  473. mediapipe/tasks/python/components/utils/cosine_similarity.py +68 -0
  474. mediapipe/tasks/python/core/__init__.py +13 -0
  475. mediapipe/tasks/python/core/base_options.py +121 -0
  476. mediapipe/tasks/python/core/optional_dependencies.py +25 -0
  477. mediapipe/tasks/python/core/task_info.py +139 -0
  478. mediapipe/tasks/python/genai/__init__.py +14 -0
  479. mediapipe/tasks/python/genai/bundler/__init__.py +23 -0
  480. mediapipe/tasks/python/genai/bundler/llm_bundler.py +130 -0
  481. mediapipe/tasks/python/genai/bundler/llm_bundler_test.py +168 -0
  482. mediapipe/tasks/python/genai/converter/__init__.py +24 -0
  483. mediapipe/tasks/python/genai/converter/converter_base.py +179 -0
  484. mediapipe/tasks/python/genai/converter/converter_factory.py +79 -0
  485. mediapipe/tasks/python/genai/converter/llm_converter.py +374 -0
  486. mediapipe/tasks/python/genai/converter/llm_converter_test.py +63 -0
  487. mediapipe/tasks/python/genai/converter/pytorch_converter.py +318 -0
  488. mediapipe/tasks/python/genai/converter/pytorch_converter_test.py +86 -0
  489. mediapipe/tasks/python/genai/converter/quantization_util.py +516 -0
  490. mediapipe/tasks/python/genai/converter/quantization_util_test.py +259 -0
  491. mediapipe/tasks/python/genai/converter/safetensors_converter.py +580 -0
  492. mediapipe/tasks/python/genai/converter/safetensors_converter_test.py +83 -0
  493. mediapipe/tasks/python/genai/converter/weight_bins_writer.py +120 -0
  494. mediapipe/tasks/python/genai/converter/weight_bins_writer_test.py +95 -0
  495. mediapipe/tasks/python/metadata/__init__.py +13 -0
  496. mediapipe/tasks/python/metadata/flatbuffers_lib/_pywrap_flatbuffers.cpython-39-x86_64-linux-gnu.so +0 -0
  497. mediapipe/tasks/python/metadata/metadata.py +928 -0
  498. mediapipe/tasks/python/metadata/metadata_displayer_cli.py +34 -0
  499. mediapipe/tasks/python/metadata/metadata_writers/__init__.py +13 -0
  500. mediapipe/tasks/python/metadata/metadata_writers/face_stylizer.py +138 -0
  501. mediapipe/tasks/python/metadata/metadata_writers/image_classifier.py +71 -0
  502. mediapipe/tasks/python/metadata/metadata_writers/image_segmenter.py +170 -0
  503. mediapipe/tasks/python/metadata/metadata_writers/metadata_info.py +1166 -0
  504. mediapipe/tasks/python/metadata/metadata_writers/metadata_writer.py +845 -0
  505. mediapipe/tasks/python/metadata/metadata_writers/model_asset_bundle_utils.py +71 -0
  506. mediapipe/tasks/python/metadata/metadata_writers/object_detector.py +331 -0
  507. mediapipe/tasks/python/metadata/metadata_writers/text_classifier.py +119 -0
  508. mediapipe/tasks/python/metadata/metadata_writers/writer_utils.py +91 -0
  509. mediapipe/tasks/python/test/__init__.py +13 -0
  510. mediapipe/tasks/python/test/audio/__init__.py +13 -0
  511. mediapipe/tasks/python/test/audio/audio_classifier_test.py +387 -0
  512. mediapipe/tasks/python/test/audio/audio_embedder_test.py +297 -0
  513. mediapipe/tasks/python/test/test_utils.py +196 -0
  514. mediapipe/tasks/python/test/text/__init__.py +13 -0
  515. mediapipe/tasks/python/test/text/language_detector_test.py +228 -0
  516. mediapipe/tasks/python/test/text/text_classifier_test.py +235 -0
  517. mediapipe/tasks/python/test/text/text_embedder_test.py +326 -0
  518. mediapipe/tasks/python/test/vision/__init__.py +13 -0
  519. mediapipe/tasks/python/test/vision/face_aligner_test.py +190 -0
  520. mediapipe/tasks/python/test/vision/face_detector_test.py +523 -0
  521. mediapipe/tasks/python/test/vision/face_landmarker_test.py +565 -0
  522. mediapipe/tasks/python/test/vision/face_stylizer_test.py +191 -0
  523. mediapipe/tasks/python/test/vision/hand_landmarker_test.py +437 -0
  524. mediapipe/tasks/python/test/vision/holistic_landmarker_test.py +544 -0
  525. mediapipe/tasks/python/test/vision/image_classifier_test.py +657 -0
  526. mediapipe/tasks/python/test/vision/image_embedder_test.py +423 -0
  527. mediapipe/tasks/python/test/vision/image_segmenter_test.py +512 -0
  528. mediapipe/tasks/python/test/vision/interactive_segmenter_test.py +341 -0
  529. mediapipe/tasks/python/test/vision/object_detector_test.py +493 -0
  530. mediapipe/tasks/python/test/vision/pose_landmarker_test.py +518 -0
  531. mediapipe/tasks/python/text/__init__.py +35 -0
  532. mediapipe/tasks/python/text/core/__init__.py +16 -0
  533. mediapipe/tasks/python/text/core/base_text_task_api.py +54 -0
  534. mediapipe/tasks/python/text/language_detector.py +220 -0
  535. mediapipe/tasks/python/text/text_classifier.py +187 -0
  536. mediapipe/tasks/python/text/text_embedder.py +188 -0
  537. mediapipe/tasks/python/vision/__init__.py +90 -0
  538. mediapipe/tasks/python/vision/core/__init__.py +14 -0
  539. mediapipe/tasks/python/vision/core/base_vision_task_api.py +226 -0
  540. mediapipe/tasks/python/vision/core/image_processing_options.py +39 -0
  541. mediapipe/tasks/python/vision/core/vision_task_running_mode.py +31 -0
  542. mediapipe/tasks/python/vision/face_aligner.py +158 -0
  543. mediapipe/tasks/python/vision/face_detector.py +332 -0
  544. mediapipe/tasks/python/vision/face_landmarker.py +3244 -0
  545. mediapipe/tasks/python/vision/face_stylizer.py +158 -0
  546. mediapipe/tasks/python/vision/gesture_recognizer.py +480 -0
  547. mediapipe/tasks/python/vision/hand_landmarker.py +504 -0
  548. mediapipe/tasks/python/vision/holistic_landmarker.py +576 -0
  549. mediapipe/tasks/python/vision/image_classifier.py +358 -0
  550. mediapipe/tasks/python/vision/image_embedder.py +362 -0
  551. mediapipe/tasks/python/vision/image_segmenter.py +433 -0
  552. mediapipe/tasks/python/vision/interactive_segmenter.py +285 -0
  553. mediapipe/tasks/python/vision/object_detector.py +389 -0
  554. mediapipe/tasks/python/vision/pose_landmarker.py +455 -0
  555. mediapipe/util/__init__.py +0 -0
  556. mediapipe/util/analytics/__init__.py +0 -0
  557. mediapipe/util/analytics/mediapipe_log_extension_pb2.py +44 -0
  558. mediapipe/util/analytics/mediapipe_logging_enums_pb2.py +37 -0
  559. mediapipe/util/audio_decoder_pb2.py +33 -0
  560. mediapipe/util/color_pb2.py +33 -0
  561. mediapipe/util/label_map_pb2.py +27 -0
  562. mediapipe/util/render_data_pb2.py +58 -0
  563. mediapipe/util/sequence/__init__.py +14 -0
  564. mediapipe/util/sequence/media_sequence.py +716 -0
  565. mediapipe/util/sequence/media_sequence_test.py +290 -0
  566. mediapipe/util/sequence/media_sequence_util.py +800 -0
  567. mediapipe/util/sequence/media_sequence_util_test.py +389 -0
  568. mediapipe/util/tracking/__init__.py +0 -0
  569. mediapipe/util/tracking/box_detector_pb2.py +39 -0
  570. mediapipe/util/tracking/box_tracker_pb2.py +32 -0
  571. mediapipe/util/tracking/camera_motion_pb2.py +31 -0
  572. mediapipe/util/tracking/flow_packager_pb2.py +60 -0
  573. mediapipe/util/tracking/frame_selection_pb2.py +35 -0
  574. mediapipe/util/tracking/frame_selection_solution_evaluator_pb2.py +28 -0
  575. mediapipe/util/tracking/motion_analysis_pb2.py +35 -0
  576. mediapipe/util/tracking/motion_estimation_pb2.py +66 -0
  577. mediapipe/util/tracking/motion_models_pb2.py +42 -0
  578. mediapipe/util/tracking/motion_saliency_pb2.py +26 -0
  579. mediapipe/util/tracking/push_pull_filtering_pb2.py +26 -0
  580. mediapipe/util/tracking/region_flow_computation_pb2.py +59 -0
  581. mediapipe/util/tracking/region_flow_pb2.py +49 -0
  582. mediapipe/util/tracking/tone_estimation_pb2.py +45 -0
  583. mediapipe/util/tracking/tone_models_pb2.py +32 -0
  584. mediapipe/util/tracking/tracked_detection_manager_config_pb2.py +26 -0
  585. mediapipe/util/tracking/tracking_pb2.py +73 -0
  586. mediapipe_nightly-0.10.21.post20241223.dist-info/LICENSE +218 -0
  587. mediapipe_nightly-0.10.21.post20241223.dist-info/METADATA +199 -0
  588. mediapipe_nightly-0.10.21.post20241223.dist-info/RECORD +593 -0
  589. mediapipe_nightly-0.10.21.post20241223.dist-info/WHEEL +5 -0
  590. mediapipe_nightly-0.10.21.post20241223.dist-info/top_level.txt +4 -0
  591. mediapipe_nightly.libs/libEGL-48f73270.so.1.1.0 +0 -0
  592. mediapipe_nightly.libs/libGLESv2-ed5eda4f.so.2.1.0 +0 -0
  593. mediapipe_nightly.libs/libGLdispatch-64b28464.so.0.0.0 +0 -0
@@ -0,0 +1,285 @@
1
+ # Copyright 2023 The MediaPipe Authors.
2
+ #
3
+ # Licensed under the Apache License, Version 2.0 (the "License");
4
+ # you may not use this file except in compliance with the License.
5
+ # You may obtain a copy of the License at
6
+ #
7
+ # http://www.apache.org/licenses/LICENSE-2.0
8
+ #
9
+ # Unless required by applicable law or agreed to in writing, software
10
+ # distributed under the License is distributed on an "AS IS" BASIS,
11
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ # See the License for the specific language governing permissions and
13
+ # limitations under the License.
14
+ """MediaPipe interactive segmenter task."""
15
+
16
+ import dataclasses
17
+ import enum
18
+ from typing import List, Optional
19
+
20
+ from mediapipe.python import packet_creator
21
+ from mediapipe.python import packet_getter
22
+ from mediapipe.python._framework_bindings import image as image_module
23
+ from mediapipe.tasks.cc.vision.image_segmenter.proto import image_segmenter_graph_options_pb2
24
+ from mediapipe.tasks.cc.vision.image_segmenter.proto import segmenter_options_pb2
25
+ from mediapipe.tasks.python.components.containers import keypoint as keypoint_module
26
+ from mediapipe.tasks.python.core import base_options as base_options_module
27
+ from mediapipe.tasks.python.core import task_info as task_info_module
28
+ from mediapipe.tasks.python.core.optional_dependencies import doc_controls
29
+ from mediapipe.tasks.python.vision.core import base_vision_task_api
30
+ from mediapipe.tasks.python.vision.core import image_processing_options as image_processing_options_module
31
+ from mediapipe.tasks.python.vision.core import vision_task_running_mode
32
+ from mediapipe.util import render_data_pb2
33
+
34
+ _BaseOptions = base_options_module.BaseOptions
35
+ _RenderDataProto = render_data_pb2.RenderData
36
+ _SegmenterOptionsProto = segmenter_options_pb2.SegmenterOptions
37
+ _ImageSegmenterGraphOptionsProto = (
38
+ image_segmenter_graph_options_pb2.ImageSegmenterGraphOptions
39
+ )
40
+ _RunningMode = vision_task_running_mode.VisionTaskRunningMode
41
+ _ImageProcessingOptions = image_processing_options_module.ImageProcessingOptions
42
+ _TaskInfo = task_info_module.TaskInfo
43
+
44
+ _CONFIDENCE_MASKS_STREAM_NAME = 'confidence_masks'
45
+ _CONFIDENCE_MASKS_TAG = 'CONFIDENCE_MASKS'
46
+ _CATEGORY_MASK_STREAM_NAME = 'category_mask'
47
+ _CATEGORY_MASK_TAG = 'CATEGORY_MASK'
48
+ _IMAGE_IN_STREAM_NAME = 'image_in'
49
+ _IMAGE_OUT_STREAM_NAME = 'image_out'
50
+ _ROI_STREAM_NAME = 'roi_in'
51
+ _ROI_TAG = 'ROI'
52
+ _NORM_RECT_STREAM_NAME = 'norm_rect_in'
53
+ _NORM_RECT_TAG = 'NORM_RECT'
54
+ _IMAGE_TAG = 'IMAGE'
55
+ _TASK_GRAPH_NAME = (
56
+ 'mediapipe.tasks.vision.interactive_segmenter.InteractiveSegmenterGraph'
57
+ )
58
+
59
+
60
+ @dataclasses.dataclass
61
+ class InteractiveSegmenterResult:
62
+ """Output result of InteractiveSegmenter.
63
+
64
+ confidence_masks: multiple masks of float image where, for each mask, each
65
+ pixel represents the prediction confidence, usually in the [0, 1] range.
66
+
67
+ category_mask: a category mask of uint8 image where each pixel represents the
68
+ class which the pixel in the original image was predicted to belong to.
69
+ """
70
+
71
+ confidence_masks: Optional[List[image_module.Image]] = None
72
+ category_mask: Optional[image_module.Image] = None
73
+
74
+
75
+ @dataclasses.dataclass
76
+ class InteractiveSegmenterOptions:
77
+ """Options for the interactive segmenter task.
78
+
79
+ Attributes:
80
+ base_options: Base options for the interactive segmenter task.
81
+ output_confidence_masks: Whether to output confidence masks.
82
+ output_category_mask: Whether to output category mask.
83
+ """
84
+
85
+ base_options: _BaseOptions
86
+ output_confidence_masks: bool = True
87
+ output_category_mask: bool = False
88
+
89
+ @doc_controls.do_not_generate_docs
90
+ def to_pb2(self) -> _ImageSegmenterGraphOptionsProto:
91
+ """Generates an ImageSegmenterGraphOptions protobuf object."""
92
+ base_options_proto = self.base_options.to_pb2()
93
+ base_options_proto.use_stream_mode = False
94
+ segmenter_options_proto = _SegmenterOptionsProto()
95
+ return _ImageSegmenterGraphOptionsProto(
96
+ base_options=base_options_proto,
97
+ segmenter_options=segmenter_options_proto,
98
+ )
99
+
100
+
101
+ @dataclasses.dataclass
102
+ class RegionOfInterest:
103
+ """The Region-Of-Interest (ROI) to interact with."""
104
+
105
+ class Format(enum.Enum):
106
+ UNSPECIFIED = 0
107
+ KEYPOINT = 1
108
+
109
+ format: Format
110
+ keypoint: Optional[keypoint_module.NormalizedKeypoint] = None
111
+
112
+
113
+ def _convert_roi_to_render_data(roi: RegionOfInterest) -> _RenderDataProto:
114
+ """Converts region of interest to render data proto."""
115
+ result = _RenderDataProto()
116
+
117
+ if roi is not None:
118
+ if roi.format == RegionOfInterest.Format.UNSPECIFIED:
119
+ raise ValueError('RegionOfInterest format not specified.')
120
+
121
+ elif roi.format == RegionOfInterest.Format.KEYPOINT:
122
+ if roi.keypoint is not None:
123
+ annotation = result.render_annotations.add()
124
+ annotation.color.r = 255
125
+ point = annotation.point
126
+ point.normalized = True
127
+ point.x = roi.keypoint.x
128
+ point.y = roi.keypoint.y
129
+ return result
130
+ else:
131
+ raise ValueError('Please specify the Region-of-interest for segmentation.')
132
+
133
+ raise ValueError('Unrecognized format.')
134
+
135
+
136
+ class InteractiveSegmenter(base_vision_task_api.BaseVisionTaskApi):
137
+ """Class that performs interactive segmentation on images.
138
+
139
+ Users can represent user interaction through `RegionOfInterest`, which gives
140
+ a hint to InteractiveSegmenter to perform segmentation focusing on the given
141
+ region of interest.
142
+
143
+ The API expects a TFLite model with mandatory TFLite Model Metadata.
144
+
145
+ Input tensor:
146
+ (kTfLiteUInt8/kTfLiteFloat32)
147
+ - image input of size `[batch x height x width x channels]`.
148
+ - batch inference is not supported (`batch` is required to be 1).
149
+ - RGB and greyscale inputs are supported (`channels` is required to be
150
+ 1 or 3).
151
+ - if type is kTfLiteFloat32, NormalizationOptions are required to be
152
+ attached to the metadata for input normalization.
153
+ Output tensors:
154
+ (kTfLiteUInt8/kTfLiteFloat32)
155
+ - list of segmented masks.
156
+ - if `output_type` is CATEGORY_MASK, uint8 Image, Image vector of size 1.
157
+ - if `output_type` is CONFIDENCE_MASK, float32 Image list of size
158
+ `channels`.
159
+ - batch is always 1
160
+
161
+ An example of such model can be found at:
162
+ https://tfhub.dev/tensorflow/lite-model/deeplabv3/1/metadata/2
163
+ """
164
+
165
+ @classmethod
166
+ def create_from_model_path(cls, model_path: str) -> 'InteractiveSegmenter':
167
+ """Creates an `InteractiveSegmenter` object from a TensorFlow Lite model and the default `InteractiveSegmenterOptions`.
168
+
169
+ Note that the created `InteractiveSegmenter` instance is in image mode, for
170
+ performing image segmentation on single image inputs.
171
+
172
+ Args:
173
+ model_path: Path to the model.
174
+
175
+ Returns:
176
+ `InteractiveSegmenter` object that's created from the model file and the
177
+ default `InteractiveSegmenterOptions`.
178
+
179
+ Raises:
180
+ ValueError: If failed to create `InteractiveSegmenter` object from the
181
+ provided file such as invalid file path.
182
+ RuntimeError: If other types of error occurred.
183
+ """
184
+ base_options = _BaseOptions(model_asset_path=model_path)
185
+ options = InteractiveSegmenterOptions(base_options=base_options)
186
+ return cls.create_from_options(options)
187
+
188
+ @classmethod
189
+ def create_from_options(
190
+ cls, options: InteractiveSegmenterOptions
191
+ ) -> 'InteractiveSegmenter':
192
+ """Creates the `InteractiveSegmenter` object from interactive segmenter options.
193
+
194
+ Args:
195
+ options: Options for the interactive segmenter task.
196
+
197
+ Returns:
198
+ `InteractiveSegmenter` object that's created from `options`.
199
+
200
+ Raises:
201
+ ValueError: If failed to create `InteractiveSegmenter` object from
202
+ `InteractiveSegmenterOptions` such as missing the model.
203
+ RuntimeError: If other types of error occurred.
204
+ """
205
+
206
+ output_streams = [
207
+ ':'.join([_IMAGE_TAG, _IMAGE_OUT_STREAM_NAME]),
208
+ ]
209
+
210
+ if options.output_confidence_masks:
211
+ output_streams.append(
212
+ ':'.join([_CONFIDENCE_MASKS_TAG, _CONFIDENCE_MASKS_STREAM_NAME])
213
+ )
214
+
215
+ if options.output_category_mask:
216
+ output_streams.append(
217
+ ':'.join([_CATEGORY_MASK_TAG, _CATEGORY_MASK_STREAM_NAME])
218
+ )
219
+
220
+ task_info = _TaskInfo(
221
+ task_graph=_TASK_GRAPH_NAME,
222
+ input_streams=[
223
+ ':'.join([_IMAGE_TAG, _IMAGE_IN_STREAM_NAME]),
224
+ ':'.join([_ROI_TAG, _ROI_STREAM_NAME]),
225
+ ':'.join([_NORM_RECT_TAG, _NORM_RECT_STREAM_NAME]),
226
+ ],
227
+ output_streams=output_streams,
228
+ task_options=options,
229
+ )
230
+ return cls(
231
+ task_info.generate_graph_config(enable_flow_limiting=False),
232
+ _RunningMode.IMAGE,
233
+ None,
234
+ )
235
+
236
+ def segment(
237
+ self,
238
+ image: image_module.Image,
239
+ roi: RegionOfInterest,
240
+ image_processing_options: Optional[_ImageProcessingOptions] = None,
241
+ ) -> InteractiveSegmenterResult:
242
+ """Performs the actual segmentation task on the provided MediaPipe Image.
243
+
244
+ The image can be of any size with format RGB.
245
+
246
+ Args:
247
+ image: MediaPipe Image.
248
+ roi: Optional user-specified region of interest for segmentation.
249
+ image_processing_options: Options for image processing.
250
+
251
+ Returns:
252
+ If the output_type is CATEGORY_MASK, the returned vector of images is
253
+ per-category segmented image mask.
254
+ If the output_type is CONFIDENCE_MASK, the returned vector of images
255
+ contains only one confidence image mask. A segmentation result object that
256
+ contains a list of segmentation masks as images.
257
+
258
+ Raises:
259
+ ValueError: If any of the input arguments is invalid.
260
+ RuntimeError: If image segmentation failed to run.
261
+ """
262
+ normalized_rect = self.convert_to_normalized_rect(
263
+ image_processing_options, image, roi_allowed=False
264
+ )
265
+ render_data_proto = _convert_roi_to_render_data(roi)
266
+ output_packets = self._process_image_data({
267
+ _IMAGE_IN_STREAM_NAME: packet_creator.create_image(image),
268
+ _ROI_STREAM_NAME: packet_creator.create_proto(render_data_proto),
269
+ _NORM_RECT_STREAM_NAME: packet_creator.create_proto(
270
+ normalized_rect.to_pb2()
271
+ ),
272
+ })
273
+ segmentation_result = InteractiveSegmenterResult()
274
+
275
+ if _CONFIDENCE_MASKS_STREAM_NAME in output_packets:
276
+ segmentation_result.confidence_masks = packet_getter.get_image_list(
277
+ output_packets[_CONFIDENCE_MASKS_STREAM_NAME]
278
+ )
279
+
280
+ if _CATEGORY_MASK_STREAM_NAME in output_packets:
281
+ segmentation_result.category_mask = packet_getter.get_image(
282
+ output_packets[_CATEGORY_MASK_STREAM_NAME]
283
+ )
284
+
285
+ return segmentation_result
@@ -0,0 +1,389 @@
1
+ # Copyright 2022 The MediaPipe Authors.
2
+ #
3
+ # Licensed under the Apache License, Version 2.0 (the "License");
4
+ # you may not use this file except in compliance with the License.
5
+ # You may obtain a copy of the License at
6
+ #
7
+ # http://www.apache.org/licenses/LICENSE-2.0
8
+ #
9
+ # Unless required by applicable law or agreed to in writing, software
10
+ # distributed under the License is distributed on an "AS IS" BASIS,
11
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ # See the License for the specific language governing permissions and
13
+ # limitations under the License.
14
+ """MediaPipe object detector task."""
15
+
16
+ import dataclasses
17
+ from typing import Callable, List, Mapping, Optional
18
+
19
+ from mediapipe.python import packet_creator
20
+ from mediapipe.python import packet_getter
21
+ from mediapipe.python._framework_bindings import image as image_module
22
+ from mediapipe.python._framework_bindings import packet as packet_module
23
+ from mediapipe.tasks.cc.vision.object_detector.proto import object_detector_options_pb2
24
+ from mediapipe.tasks.python.components.containers import detections as detections_module
25
+ from mediapipe.tasks.python.components.containers import rect
26
+ from mediapipe.tasks.python.core import base_options as base_options_module
27
+ from mediapipe.tasks.python.core import task_info as task_info_module
28
+ from mediapipe.tasks.python.core.optional_dependencies import doc_controls
29
+ from mediapipe.tasks.python.vision.core import base_vision_task_api
30
+ from mediapipe.tasks.python.vision.core import image_processing_options as image_processing_options_module
31
+ from mediapipe.tasks.python.vision.core import vision_task_running_mode as running_mode_module
32
+
33
+ ObjectDetectorResult = detections_module.DetectionResult
34
+ _NormalizedRect = rect.NormalizedRect
35
+ _BaseOptions = base_options_module.BaseOptions
36
+ _ObjectDetectorOptionsProto = object_detector_options_pb2.ObjectDetectorOptions
37
+ _RunningMode = running_mode_module.VisionTaskRunningMode
38
+ _ImageProcessingOptions = image_processing_options_module.ImageProcessingOptions
39
+ _TaskInfo = task_info_module.TaskInfo
40
+
41
+ _DETECTIONS_OUT_STREAM_NAME = 'detections_out'
42
+ _DETECTIONS_TAG = 'DETECTIONS'
43
+ _IMAGE_IN_STREAM_NAME = 'image_in'
44
+ _IMAGE_OUT_STREAM_NAME = 'image_out'
45
+ _IMAGE_TAG = 'IMAGE'
46
+ _NORM_RECT_STREAM_NAME = 'norm_rect_in'
47
+ _NORM_RECT_TAG = 'NORM_RECT'
48
+ _TASK_GRAPH_NAME = 'mediapipe.tasks.vision.ObjectDetectorGraph'
49
+ _MICRO_SECONDS_PER_MILLISECOND = 1000
50
+
51
+
52
+ @dataclasses.dataclass
53
+ class ObjectDetectorOptions:
54
+ """Options for the object detector task.
55
+
56
+ Attributes:
57
+ base_options: Base options for the object detector task.
58
+ running_mode: The running mode of the task. Default to the image mode.
59
+ Object detector task has three running modes: 1) The image mode for
60
+ detecting objects on single image inputs. 2) The video mode for detecting
61
+ objects on the decoded frames of a video. 3) The live stream mode for
62
+ detecting objects on a live stream of input data, such as from camera.
63
+ display_names_locale: The locale to use for display names specified through
64
+ the TFLite Model Metadata.
65
+ max_results: The maximum number of top-scored classification results to
66
+ return.
67
+ score_threshold: Overrides the ones provided in the model metadata. Results
68
+ below this value are rejected.
69
+ category_allowlist: Allowlist of category names. If non-empty, detection
70
+ results whose category name is not in this set will be filtered out.
71
+ Duplicate or unknown category names are ignored. Mutually exclusive with
72
+ `category_denylist`.
73
+ category_denylist: Denylist of category names. If non-empty, detection
74
+ results whose category name is in this set will be filtered out. Duplicate
75
+ or unknown category names are ignored. Mutually exclusive with
76
+ `category_allowlist`.
77
+ result_callback: The user-defined result callback for processing live stream
78
+ data. The result callback should only be specified when the running mode
79
+ is set to the live stream mode.
80
+ """
81
+
82
+ base_options: _BaseOptions
83
+ running_mode: _RunningMode = _RunningMode.IMAGE
84
+ display_names_locale: Optional[str] = None
85
+ max_results: Optional[int] = None
86
+ score_threshold: Optional[float] = None
87
+ category_allowlist: Optional[List[str]] = None
88
+ category_denylist: Optional[List[str]] = None
89
+ result_callback: Optional[
90
+ Callable[[ObjectDetectorResult, image_module.Image, int], None]
91
+ ] = None
92
+
93
+ @doc_controls.do_not_generate_docs
94
+ def to_pb2(self) -> _ObjectDetectorOptionsProto:
95
+ """Generates an ObjectDetectorOptions protobuf object."""
96
+ base_options_proto = self.base_options.to_pb2()
97
+ base_options_proto.use_stream_mode = (
98
+ False if self.running_mode == _RunningMode.IMAGE else True
99
+ )
100
+ return _ObjectDetectorOptionsProto(
101
+ base_options=base_options_proto,
102
+ display_names_locale=self.display_names_locale,
103
+ max_results=self.max_results,
104
+ score_threshold=self.score_threshold,
105
+ category_allowlist=self.category_allowlist,
106
+ category_denylist=self.category_denylist,
107
+ )
108
+
109
+
110
+ class ObjectDetector(base_vision_task_api.BaseVisionTaskApi):
111
+ """Class that performs object detection on images.
112
+
113
+ The API expects a TFLite model with mandatory TFLite Model Metadata.
114
+
115
+ Input tensor:
116
+ (kTfLiteUInt8/kTfLiteFloat32)
117
+ - image input of size `[batch x height x width x channels]`.
118
+ - batch inference is not supported (`batch` is required to be 1).
119
+ - only RGB inputs are supported (`channels` is required to be 3).
120
+ - if type is kTfLiteFloat32, NormalizationOptions are required to be
121
+ attached to the metadata for input normalization.
122
+ Output tensors must be the 4 outputs of a `DetectionPostProcess` op, i.e:
123
+ (kTfLiteFloat32)
124
+ - locations tensor of size `[num_results x 4]`, the inner array
125
+ representing bounding boxes in the form [top, left, right, bottom].
126
+ - BoundingBoxProperties are required to be attached to the metadata
127
+ and must specify type=BOUNDARIES and coordinate_type=RATIO.
128
+ (kTfLiteFloat32)
129
+ - classes tensor of size `[num_results]`, each value representing the
130
+ integer index of a class.
131
+ - optional (but recommended) label map(s) can be attached as
132
+ AssociatedFile-s with type TENSOR_VALUE_LABELS, containing one label per
133
+ line. The first such AssociatedFile (if any) is used to fill the
134
+ `class_name` field of the results. The `display_name` field is filled
135
+ from the AssociatedFile (if any) whose locale matches the
136
+ `display_names_locale` field of the `ObjectDetectorOptions` used at
137
+ creation time ("en" by default, i.e. English). If none of these are
138
+ available, only the `index` field of the results will be filled.
139
+ (kTfLiteFloat32)
140
+ - scores tensor of size `[num_results]`, each value representing the score
141
+ of the detected object.
142
+ - optional score calibration can be attached using ScoreCalibrationOptions
143
+ and an AssociatedFile with type TENSOR_AXIS_SCORE_CALIBRATION. See
144
+ metadata_schema.fbs [1] for more details.
145
+ (kTfLiteFloat32)
146
+ - integer num_results as a tensor of size `[1]`
147
+
148
+ An example of such model can be found at:
149
+ https://tfhub.dev/google/lite-model/object_detection/mobile_object_localizer_v1/1/metadata/1
150
+
151
+ [1]:
152
+ https://github.com/google/mediapipe/blob/6cdc6443b6a7ed662744e2a2ce2d58d9c83e6d6f/mediapipe/tasks/metadata/metadata_schema.fbs#L456
153
+ """
154
+
155
+ @classmethod
156
+ def create_from_model_path(cls, model_path: str) -> 'ObjectDetector':
157
+ """Creates an `ObjectDetector` object from a TensorFlow Lite model and the default `ObjectDetectorOptions`.
158
+
159
+ Note that the created `ObjectDetector` instance is in image mode, for
160
+ detecting objects on single image inputs.
161
+
162
+ Args:
163
+ model_path: Path to the model.
164
+
165
+ Returns:
166
+ `ObjectDetector` object that's created from the model file and the default
167
+ `ObjectDetectorOptions`.
168
+
169
+ Raises:
170
+ ValueError: If failed to create `ObjectDetector` object from the provided
171
+ file such as invalid file path.
172
+ RuntimeError: If other types of error occurred.
173
+ """
174
+ base_options = _BaseOptions(model_asset_path=model_path)
175
+ options = ObjectDetectorOptions(
176
+ base_options=base_options, running_mode=_RunningMode.IMAGE
177
+ )
178
+ return cls.create_from_options(options)
179
+
180
+ @classmethod
181
+ def create_from_options(
182
+ cls, options: ObjectDetectorOptions
183
+ ) -> 'ObjectDetector':
184
+ """Creates the `ObjectDetector` object from object detector options.
185
+
186
+ Args:
187
+ options: Options for the object detector task.
188
+
189
+ Returns:
190
+ `ObjectDetector` object that's created from `options`.
191
+
192
+ Raises:
193
+ ValueError: If failed to create `ObjectDetector` object from
194
+ `ObjectDetectorOptions` such as missing the model.
195
+ RuntimeError: If other types of error occurred.
196
+ """
197
+
198
+ def packets_callback(output_packets: Mapping[str, packet_module.Packet]):
199
+ if output_packets[_IMAGE_OUT_STREAM_NAME].is_empty():
200
+ return
201
+ image = packet_getter.get_image(output_packets[_IMAGE_OUT_STREAM_NAME])
202
+ if output_packets[_DETECTIONS_OUT_STREAM_NAME].is_empty():
203
+ empty_packet = output_packets[_DETECTIONS_OUT_STREAM_NAME]
204
+ options.result_callback(
205
+ ObjectDetectorResult([]),
206
+ image,
207
+ empty_packet.timestamp.value // _MICRO_SECONDS_PER_MILLISECOND,
208
+ )
209
+ return
210
+ detection_proto_list = packet_getter.get_proto_list(
211
+ output_packets[_DETECTIONS_OUT_STREAM_NAME]
212
+ )
213
+ detection_result = ObjectDetectorResult(
214
+ [
215
+ detections_module.Detection.create_from_pb2(result)
216
+ for result in detection_proto_list
217
+ ]
218
+ )
219
+ timestamp = output_packets[_IMAGE_OUT_STREAM_NAME].timestamp
220
+ options.result_callback(
221
+ detection_result,
222
+ image,
223
+ timestamp.value // _MICRO_SECONDS_PER_MILLISECOND,
224
+ )
225
+
226
+ task_info = _TaskInfo(
227
+ task_graph=_TASK_GRAPH_NAME,
228
+ input_streams=[
229
+ ':'.join([_IMAGE_TAG, _IMAGE_IN_STREAM_NAME]),
230
+ ':'.join([_NORM_RECT_TAG, _NORM_RECT_STREAM_NAME]),
231
+ ],
232
+ output_streams=[
233
+ ':'.join([_DETECTIONS_TAG, _DETECTIONS_OUT_STREAM_NAME]),
234
+ ':'.join([_IMAGE_TAG, _IMAGE_OUT_STREAM_NAME]),
235
+ ],
236
+ task_options=options,
237
+ )
238
+ return cls(
239
+ task_info.generate_graph_config(
240
+ enable_flow_limiting=options.running_mode
241
+ == _RunningMode.LIVE_STREAM
242
+ ),
243
+ options.running_mode,
244
+ packets_callback if options.result_callback else None,
245
+ )
246
+
247
+ # TODO: Create an Image class for MediaPipe Tasks.
248
+ def detect(
249
+ self,
250
+ image: image_module.Image,
251
+ image_processing_options: Optional[_ImageProcessingOptions] = None,
252
+ ) -> ObjectDetectorResult:
253
+ """Performs object detection on the provided MediaPipe Image.
254
+
255
+ Only use this method when the ObjectDetector is created with the image
256
+ running mode.
257
+
258
+ Args:
259
+ image: MediaPipe Image.
260
+ image_processing_options: Options for image processing.
261
+
262
+ Returns:
263
+ A detection result object that contains a list of detections, each
264
+ detection has a bounding box that is expressed in the unrotated input
265
+ frame of reference coordinates system, i.e. in `[0,image_width) x [0,
266
+ image_height)`, which are the dimensions of the underlying image data.
267
+
268
+ Raises:
269
+ ValueError: If any of the input arguments is invalid.
270
+ RuntimeError: If object detection failed to run.
271
+ """
272
+ normalized_rect = self.convert_to_normalized_rect(
273
+ image_processing_options, image, roi_allowed=False
274
+ )
275
+ output_packets = self._process_image_data({
276
+ _IMAGE_IN_STREAM_NAME: packet_creator.create_image(image),
277
+ _NORM_RECT_STREAM_NAME: packet_creator.create_proto(
278
+ normalized_rect.to_pb2()
279
+ ),
280
+ })
281
+ if output_packets[_DETECTIONS_OUT_STREAM_NAME].is_empty():
282
+ return ObjectDetectorResult([])
283
+ detection_proto_list = packet_getter.get_proto_list(
284
+ output_packets[_DETECTIONS_OUT_STREAM_NAME]
285
+ )
286
+ return ObjectDetectorResult(
287
+ [
288
+ detections_module.Detection.create_from_pb2(result)
289
+ for result in detection_proto_list
290
+ ]
291
+ )
292
+
293
+ def detect_for_video(
294
+ self,
295
+ image: image_module.Image,
296
+ timestamp_ms: int,
297
+ image_processing_options: Optional[_ImageProcessingOptions] = None,
298
+ ) -> ObjectDetectorResult:
299
+ """Performs object detection on the provided video frames.
300
+
301
+ Only use this method when the ObjectDetector is created with the video
302
+ running mode. It's required to provide the video frame's timestamp (in
303
+ milliseconds) along with the video frame. The input timestamps should be
304
+ monotonically increasing for adjacent calls of this method.
305
+
306
+ Args:
307
+ image: MediaPipe Image.
308
+ timestamp_ms: The timestamp of the input video frame in milliseconds.
309
+ image_processing_options: Options for image processing.
310
+
311
+ Returns:
312
+ A detection result object that contains a list of detections, each
313
+ detection has a bounding box that is expressed in the unrotated input
314
+ frame of reference coordinates system, i.e. in `[0,image_width) x [0,
315
+ image_height)`, which are the dimensions of the underlying image data.
316
+
317
+ Raises:
318
+ ValueError: If any of the input arguments is invalid.
319
+ RuntimeError: If object detection failed to run.
320
+ """
321
+ normalized_rect = self.convert_to_normalized_rect(
322
+ image_processing_options, image, roi_allowed=False
323
+ )
324
+ output_packets = self._process_video_data({
325
+ _IMAGE_IN_STREAM_NAME: packet_creator.create_image(image).at(
326
+ timestamp_ms * _MICRO_SECONDS_PER_MILLISECOND
327
+ ),
328
+ _NORM_RECT_STREAM_NAME: packet_creator.create_proto(
329
+ normalized_rect.to_pb2()
330
+ ).at(timestamp_ms * _MICRO_SECONDS_PER_MILLISECOND),
331
+ })
332
+ if output_packets[_DETECTIONS_OUT_STREAM_NAME].is_empty():
333
+ return ObjectDetectorResult([])
334
+ detection_proto_list = packet_getter.get_proto_list(
335
+ output_packets[_DETECTIONS_OUT_STREAM_NAME]
336
+ )
337
+ return ObjectDetectorResult(
338
+ [
339
+ detections_module.Detection.create_from_pb2(result)
340
+ for result in detection_proto_list
341
+ ]
342
+ )
343
+
344
+ def detect_async(
345
+ self,
346
+ image: image_module.Image,
347
+ timestamp_ms: int,
348
+ image_processing_options: Optional[_ImageProcessingOptions] = None,
349
+ ) -> None:
350
+ """Sends live image data (an Image with a unique timestamp) to perform object detection.
351
+
352
+ Only use this method when the ObjectDetector is created with the live stream
353
+ running mode. The input timestamps should be monotonically increasing for
354
+ adjacent calls of this method. This method will return immediately after the
355
+ input image is accepted. The results will be available via the
356
+ `result_callback` provided in the `ObjectDetectorOptions`. The
357
+ `detect_async` method is designed to process live stream data such as camera
358
+ input. To lower the overall latency, object detector may drop the input
359
+ images if needed. In other words, it's not guaranteed to have output per
360
+ input image.
361
+
362
+ The `result_callback` prvoides:
363
+ - A detection result object that contains a list of detections, each
364
+ detection has a bounding box that is expressed in the unrotated input
365
+ frame of reference coordinates system, i.e. in `[0,image_width) x [0,
366
+ image_height)`, which are the dimensions of the underlying image data.
367
+ - The input image that the object detector runs on.
368
+ - The input timestamp in milliseconds.
369
+
370
+ Args:
371
+ image: MediaPipe Image.
372
+ timestamp_ms: The timestamp of the input image in milliseconds.
373
+ image_processing_options: Options for image processing.
374
+
375
+ Raises:
376
+ ValueError: If the current input timestamp is smaller than what the object
377
+ detector has already processed.
378
+ """
379
+ normalized_rect = self.convert_to_normalized_rect(
380
+ image_processing_options, image, roi_allowed=False
381
+ )
382
+ self._send_live_stream_data({
383
+ _IMAGE_IN_STREAM_NAME: packet_creator.create_image(image).at(
384
+ timestamp_ms * _MICRO_SECONDS_PER_MILLISECOND
385
+ ),
386
+ _NORM_RECT_STREAM_NAME: packet_creator.create_proto(
387
+ normalized_rect.to_pb2()
388
+ ).at(timestamp_ms * _MICRO_SECONDS_PER_MILLISECOND),
389
+ })