inference-models 0.18.3__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (195) hide show
  1. inference_models/__init__.py +36 -0
  2. inference_models/configuration.py +72 -0
  3. inference_models/constants.py +2 -0
  4. inference_models/entities.py +5 -0
  5. inference_models/errors.py +137 -0
  6. inference_models/logger.py +52 -0
  7. inference_models/model_pipelines/__init__.py +0 -0
  8. inference_models/model_pipelines/auto_loaders/__init__.py +0 -0
  9. inference_models/model_pipelines/auto_loaders/core.py +120 -0
  10. inference_models/model_pipelines/auto_loaders/pipelines_registry.py +36 -0
  11. inference_models/model_pipelines/face_and_gaze_detection/__init__.py +0 -0
  12. inference_models/model_pipelines/face_and_gaze_detection/mediapipe_l2cs.py +200 -0
  13. inference_models/models/__init__.py +0 -0
  14. inference_models/models/auto_loaders/__init__.py +0 -0
  15. inference_models/models/auto_loaders/access_manager.py +168 -0
  16. inference_models/models/auto_loaders/auto_negotiation.py +1329 -0
  17. inference_models/models/auto_loaders/auto_resolution_cache.py +129 -0
  18. inference_models/models/auto_loaders/constants.py +7 -0
  19. inference_models/models/auto_loaders/core.py +1341 -0
  20. inference_models/models/auto_loaders/dependency_models.py +52 -0
  21. inference_models/models/auto_loaders/entities.py +57 -0
  22. inference_models/models/auto_loaders/models_registry.py +497 -0
  23. inference_models/models/auto_loaders/presentation_utils.py +333 -0
  24. inference_models/models/auto_loaders/ranking.py +413 -0
  25. inference_models/models/auto_loaders/utils.py +31 -0
  26. inference_models/models/base/__init__.py +0 -0
  27. inference_models/models/base/classification.py +123 -0
  28. inference_models/models/base/depth_estimation.py +62 -0
  29. inference_models/models/base/documents_parsing.py +111 -0
  30. inference_models/models/base/embeddings.py +66 -0
  31. inference_models/models/base/instance_segmentation.py +87 -0
  32. inference_models/models/base/keypoints_detection.py +93 -0
  33. inference_models/models/base/object_detection.py +143 -0
  34. inference_models/models/base/semantic_segmentation.py +74 -0
  35. inference_models/models/base/types.py +5 -0
  36. inference_models/models/clip/__init__.py +0 -0
  37. inference_models/models/clip/clip_onnx.py +148 -0
  38. inference_models/models/clip/clip_pytorch.py +104 -0
  39. inference_models/models/clip/preprocessing.py +162 -0
  40. inference_models/models/common/__init__.py +0 -0
  41. inference_models/models/common/cuda.py +30 -0
  42. inference_models/models/common/model_packages.py +25 -0
  43. inference_models/models/common/onnx.py +379 -0
  44. inference_models/models/common/roboflow/__init__.py +0 -0
  45. inference_models/models/common/roboflow/model_packages.py +361 -0
  46. inference_models/models/common/roboflow/post_processing.py +436 -0
  47. inference_models/models/common/roboflow/pre_processing.py +1332 -0
  48. inference_models/models/common/torch.py +20 -0
  49. inference_models/models/common/trt.py +266 -0
  50. inference_models/models/deep_lab_v3_plus/__init__.py +0 -0
  51. inference_models/models/deep_lab_v3_plus/deep_lab_v3_plus_segmentation_onnx.py +282 -0
  52. inference_models/models/deep_lab_v3_plus/deep_lab_v3_plus_segmentation_torch.py +264 -0
  53. inference_models/models/deep_lab_v3_plus/deep_lab_v3_plus_segmentation_trt.py +313 -0
  54. inference_models/models/depth_anything_v2/__init__.py +0 -0
  55. inference_models/models/depth_anything_v2/depth_anything_v2_hf.py +77 -0
  56. inference_models/models/dinov3/__init__.py +0 -0
  57. inference_models/models/dinov3/dinov3_classification_onnx.py +348 -0
  58. inference_models/models/dinov3/dinov3_classification_torch.py +323 -0
  59. inference_models/models/doctr/__init__.py +0 -0
  60. inference_models/models/doctr/doctr_torch.py +304 -0
  61. inference_models/models/easy_ocr/__init__.py +0 -0
  62. inference_models/models/easy_ocr/easy_ocr_torch.py +222 -0
  63. inference_models/models/florence2/__init__.py +0 -0
  64. inference_models/models/florence2/florence2_hf.py +897 -0
  65. inference_models/models/grounding_dino/__init__.py +0 -0
  66. inference_models/models/grounding_dino/grounding_dino_torch.py +227 -0
  67. inference_models/models/l2cs/__init__.py +0 -0
  68. inference_models/models/l2cs/l2cs_onnx.py +216 -0
  69. inference_models/models/mediapipe_face_detection/__init__.py +0 -0
  70. inference_models/models/mediapipe_face_detection/face_detection.py +203 -0
  71. inference_models/models/moondream2/__init__.py +0 -0
  72. inference_models/models/moondream2/moondream2_hf.py +281 -0
  73. inference_models/models/owlv2/__init__.py +0 -0
  74. inference_models/models/owlv2/cache.py +182 -0
  75. inference_models/models/owlv2/entities.py +112 -0
  76. inference_models/models/owlv2/owlv2_hf.py +695 -0
  77. inference_models/models/owlv2/reference_dataset.py +291 -0
  78. inference_models/models/paligemma/__init__.py +0 -0
  79. inference_models/models/paligemma/paligemma_hf.py +209 -0
  80. inference_models/models/perception_encoder/__init__.py +0 -0
  81. inference_models/models/perception_encoder/perception_encoder_pytorch.py +197 -0
  82. inference_models/models/perception_encoder/vision_encoder/__init__.py +0 -0
  83. inference_models/models/perception_encoder/vision_encoder/config.py +160 -0
  84. inference_models/models/perception_encoder/vision_encoder/pe.py +742 -0
  85. inference_models/models/perception_encoder/vision_encoder/rope.py +344 -0
  86. inference_models/models/perception_encoder/vision_encoder/tokenizer.py +342 -0
  87. inference_models/models/perception_encoder/vision_encoder/transforms.py +33 -0
  88. inference_models/models/qwen25vl/__init__.py +1 -0
  89. inference_models/models/qwen25vl/qwen25vl_hf.py +285 -0
  90. inference_models/models/resnet/__init__.py +0 -0
  91. inference_models/models/resnet/resnet_classification_onnx.py +330 -0
  92. inference_models/models/resnet/resnet_classification_torch.py +305 -0
  93. inference_models/models/resnet/resnet_classification_trt.py +369 -0
  94. inference_models/models/rfdetr/__init__.py +0 -0
  95. inference_models/models/rfdetr/backbone_builder.py +101 -0
  96. inference_models/models/rfdetr/class_remapping.py +41 -0
  97. inference_models/models/rfdetr/common.py +115 -0
  98. inference_models/models/rfdetr/default_labels.py +108 -0
  99. inference_models/models/rfdetr/dinov2_with_windowed_attn.py +1330 -0
  100. inference_models/models/rfdetr/misc.py +26 -0
  101. inference_models/models/rfdetr/ms_deform_attn.py +180 -0
  102. inference_models/models/rfdetr/ms_deform_attn_func.py +60 -0
  103. inference_models/models/rfdetr/position_encoding.py +166 -0
  104. inference_models/models/rfdetr/post_processor.py +83 -0
  105. inference_models/models/rfdetr/projector.py +373 -0
  106. inference_models/models/rfdetr/rfdetr_backbone_pytorch.py +394 -0
  107. inference_models/models/rfdetr/rfdetr_base_pytorch.py +807 -0
  108. inference_models/models/rfdetr/rfdetr_instance_segmentation_onnx.py +206 -0
  109. inference_models/models/rfdetr/rfdetr_instance_segmentation_pytorch.py +373 -0
  110. inference_models/models/rfdetr/rfdetr_instance_segmentation_trt.py +227 -0
  111. inference_models/models/rfdetr/rfdetr_object_detection_onnx.py +244 -0
  112. inference_models/models/rfdetr/rfdetr_object_detection_pytorch.py +470 -0
  113. inference_models/models/rfdetr/rfdetr_object_detection_trt.py +270 -0
  114. inference_models/models/rfdetr/segmentation_head.py +273 -0
  115. inference_models/models/rfdetr/transformer.py +767 -0
  116. inference_models/models/roboflow_instant/__init__.py +0 -0
  117. inference_models/models/roboflow_instant/roboflow_instant_hf.py +141 -0
  118. inference_models/models/sam/__init__.py +0 -0
  119. inference_models/models/sam/cache.py +147 -0
  120. inference_models/models/sam/entities.py +25 -0
  121. inference_models/models/sam/sam_torch.py +675 -0
  122. inference_models/models/sam2/__init__.py +0 -0
  123. inference_models/models/sam2/cache.py +162 -0
  124. inference_models/models/sam2/entities.py +43 -0
  125. inference_models/models/sam2/sam2_torch.py +905 -0
  126. inference_models/models/sam2_rt/__init__.py +0 -0
  127. inference_models/models/sam2_rt/sam2_pytorch.py +119 -0
  128. inference_models/models/smolvlm/__init__.py +0 -0
  129. inference_models/models/smolvlm/smolvlm_hf.py +245 -0
  130. inference_models/models/trocr/__init__.py +0 -0
  131. inference_models/models/trocr/trocr_hf.py +53 -0
  132. inference_models/models/vit/__init__.py +0 -0
  133. inference_models/models/vit/vit_classification_huggingface.py +319 -0
  134. inference_models/models/vit/vit_classification_onnx.py +326 -0
  135. inference_models/models/vit/vit_classification_trt.py +365 -0
  136. inference_models/models/yolact/__init__.py +1 -0
  137. inference_models/models/yolact/yolact_instance_segmentation_onnx.py +336 -0
  138. inference_models/models/yolact/yolact_instance_segmentation_trt.py +361 -0
  139. inference_models/models/yolo_world/__init__.py +1 -0
  140. inference_models/models/yolonas/__init__.py +0 -0
  141. inference_models/models/yolonas/nms.py +44 -0
  142. inference_models/models/yolonas/yolonas_object_detection_onnx.py +204 -0
  143. inference_models/models/yolonas/yolonas_object_detection_trt.py +230 -0
  144. inference_models/models/yolov10/__init__.py +0 -0
  145. inference_models/models/yolov10/yolov10_object_detection_onnx.py +187 -0
  146. inference_models/models/yolov10/yolov10_object_detection_trt.py +215 -0
  147. inference_models/models/yolov11/__init__.py +0 -0
  148. inference_models/models/yolov11/yolov11_onnx.py +28 -0
  149. inference_models/models/yolov11/yolov11_torch_script.py +25 -0
  150. inference_models/models/yolov11/yolov11_trt.py +21 -0
  151. inference_models/models/yolov12/__init__.py +0 -0
  152. inference_models/models/yolov12/yolov12_onnx.py +7 -0
  153. inference_models/models/yolov12/yolov12_torch_script.py +7 -0
  154. inference_models/models/yolov12/yolov12_trt.py +7 -0
  155. inference_models/models/yolov5/__init__.py +0 -0
  156. inference_models/models/yolov5/nms.py +99 -0
  157. inference_models/models/yolov5/yolov5_instance_segmentation_onnx.py +225 -0
  158. inference_models/models/yolov5/yolov5_instance_segmentation_trt.py +255 -0
  159. inference_models/models/yolov5/yolov5_object_detection_onnx.py +192 -0
  160. inference_models/models/yolov5/yolov5_object_detection_trt.py +218 -0
  161. inference_models/models/yolov7/__init__.py +0 -0
  162. inference_models/models/yolov7/yolov7_instance_segmentation_onnx.py +226 -0
  163. inference_models/models/yolov7/yolov7_instance_segmentation_trt.py +253 -0
  164. inference_models/models/yolov8/__init__.py +0 -0
  165. inference_models/models/yolov8/yolov8_classification_onnx.py +181 -0
  166. inference_models/models/yolov8/yolov8_instance_segmentation_onnx.py +239 -0
  167. inference_models/models/yolov8/yolov8_instance_segmentation_torch_script.py +201 -0
  168. inference_models/models/yolov8/yolov8_instance_segmentation_trt.py +268 -0
  169. inference_models/models/yolov8/yolov8_key_points_detection_onnx.py +263 -0
  170. inference_models/models/yolov8/yolov8_key_points_detection_torch_script.py +218 -0
  171. inference_models/models/yolov8/yolov8_key_points_detection_trt.py +287 -0
  172. inference_models/models/yolov8/yolov8_object_detection_onnx.py +213 -0
  173. inference_models/models/yolov8/yolov8_object_detection_torch_script.py +166 -0
  174. inference_models/models/yolov8/yolov8_object_detection_trt.py +231 -0
  175. inference_models/models/yolov9/__init__.py +0 -0
  176. inference_models/models/yolov9/yolov9_onnx.py +7 -0
  177. inference_models/models/yolov9/yolov9_torch_script.py +7 -0
  178. inference_models/models/yolov9/yolov9_trt.py +7 -0
  179. inference_models/runtime_introspection/__init__.py +0 -0
  180. inference_models/runtime_introspection/core.py +410 -0
  181. inference_models/utils/__init__.py +0 -0
  182. inference_models/utils/download.py +608 -0
  183. inference_models/utils/environment.py +28 -0
  184. inference_models/utils/file_system.py +51 -0
  185. inference_models/utils/hashing.py +7 -0
  186. inference_models/utils/imports.py +48 -0
  187. inference_models/utils/onnx_introspection.py +17 -0
  188. inference_models/weights_providers/__init__.py +0 -0
  189. inference_models/weights_providers/core.py +20 -0
  190. inference_models/weights_providers/entities.py +159 -0
  191. inference_models/weights_providers/roboflow.py +601 -0
  192. inference_models-0.18.3.dist-info/METADATA +466 -0
  193. inference_models-0.18.3.dist-info/RECORD +195 -0
  194. inference_models-0.18.3.dist-info/WHEEL +5 -0
  195. inference_models-0.18.3.dist-info/top_level.txt +1 -0
@@ -0,0 +1,195 @@
1
+ inference_models/__init__.py,sha256=VQl38l8GyUdRb0H0tB8eYrv_8831HmNLsNtOLCY30Uc,1339
2
+ inference_models/configuration.py,sha256=lYkdnzLAYwVzaGUDnRaNYWKqBVurq52zz4ASfQQWoEw,2630
3
+ inference_models/constants.py,sha256=644eUneU8TamCzWZ_TJLIv2pPeaHYcFGgXSSNYS2yck,79
4
+ inference_models/entities.py,sha256=xhFqW0z-bWLITIjZEPMStJgw7i97z5STMgQm1MRvK4Y,168
5
+ inference_models/errors.py,sha256=mXGurGA4q3PRjwStm05EZtJP4xCBrYoOFiTHvSQfDUk,2469
6
+ inference_models/logger.py,sha256=pJR922npIgD_cjU6ePYrDMx6_QvzArB5VjM-MDLdv8U,1368
7
+ inference_models/model_pipelines/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
8
+ inference_models/model_pipelines/auto_loaders/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
9
+ inference_models/model_pipelines/auto_loaders/core.py,sha256=8Mnd5wODg_3cqxgMdP_hAEkPecNo9_99-VCtcfhTm-w,5773
10
+ inference_models/model_pipelines/auto_loaders/pipelines_registry.py,sha256=33PIoVd74t9dnKizs3VhIkLgZx-MLfFTKe6LOl1Q9cY,1290
11
+ inference_models/model_pipelines/face_and_gaze_detection/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
12
+ inference_models/model_pipelines/face_and_gaze_detection/mediapipe_l2cs.py,sha256=4bEnp0KunSOihQK9Xc4lqnRJVRtkCcCpVV5-ldBhY-E,8128
13
+ inference_models/models/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
14
+ inference_models/models/auto_loaders/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
15
+ inference_models/models/auto_loaders/access_manager.py,sha256=Se3pCczKCOTzKBIxVmjK8eiifUgp8lipg7FtM_t0RUY,4196
16
+ inference_models/models/auto_loaders/auto_negotiation.py,sha256=zLjs6G_Bnx0K0e3ls6P6KPrRw3oTgRZTm3bTgXraM4U,60006
17
+ inference_models/models/auto_loaders/auto_resolution_cache.py,sha256=wf6pQXeKIMdkiSEyphQSpeS4EHOeUYLS1r_zXn3sRvo,4836
18
+ inference_models/models/auto_loaders/constants.py,sha256=6-9jnthGIIFvWvFlext-RZmLNKDd-TXLMe6AWsLvw40,306
19
+ inference_models/models/auto_loaders/core.py,sha256=FG9QVIlXkSS9fHkZUnwYjMgJrC4Vqdq7aCFRqmTExVU,64046
20
+ inference_models/models/auto_loaders/dependency_models.py,sha256=PEU3SySjqSHPmwNKEbYihvXum_UtOQjGZ40V0M95Apw,2375
21
+ inference_models/models/auto_loaders/entities.py,sha256=GX7MvMK1geWA2UvP9GR8nUMcazB9IL33xEpwV6xW_JQ,1714
22
+ inference_models/models/auto_loaders/models_registry.py,sha256=xZySoxVkY2i2ewmF3ATnd4aRLYZsZroXuxoTxWbvfU4,22171
23
+ inference_models/models/auto_loaders/presentation_utils.py,sha256=8PXs_tBOEiD_8V6qu0L3a00wc-LZVolKhTnu-1EsDxs,13155
24
+ inference_models/models/auto_loaders/ranking.py,sha256=V846HJ33iycZr_47JOcLW6YCwMoSHhCzCKK2Q_EiXao,16157
25
+ inference_models/models/auto_loaders/utils.py,sha256=mzFl2XGQPXSpEUYgy1pKynwvBbkH3Y3zUwJbfbVUsug,1482
26
+ inference_models/models/base/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
27
+ inference_models/models/base/classification.py,sha256=Kh3DNNySyHxrnUGBv85a2rOIuqaPcBWfm_zDjgCvF_U,3484
28
+ inference_models/models/base/depth_estimation.py,sha256=3pQqIVx2JHGXujj_iGYMt8hKL7cSscn3uHw1tvUdK-0,1692
29
+ inference_models/models/base/documents_parsing.py,sha256=aMnDWd921MwziU-4G7YwWJ5e4UNMN2wQWTsTca2tekg,3130
30
+ inference_models/models/base/embeddings.py,sha256=mAGfHfx9cyD0Nig0WMkRaxAL0lyp80n4yVIR1RzmXns,1761
31
+ inference_models/models/base/instance_segmentation.py,sha256=gnWUKMFmSpWqrdjoHUMxAMIrltkAbu1p9QX-GyGPEqg,2482
32
+ inference_models/models/base/keypoints_detection.py,sha256=15Zs6ZKFlbZ3KpQZ75q6ndiPI2OJXt0hUYaObbeNrdk,2712
33
+ inference_models/models/base/object_detection.py,sha256=V-K_N80uwxDmvjYqoTKN74oYqLCyBeiUrcKnWCNp-Ps,4009
34
+ inference_models/models/base/semantic_segmentation.py,sha256=dV9ChC-yfHKNdnGUaG72JOF4Q8AJAB8t623_nIC9hko,1974
35
+ inference_models/models/base/types.py,sha256=2J95RiAwja_flSDr1ldZccqwqn0QQW2c60FNWms54Ag,177
36
+ inference_models/models/clip/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
37
+ inference_models/models/clip/clip_onnx.py,sha256=W4cXltU70wMIptiZsr7CX4xzipxZkBRv9rJGIucLkZ4,6205
38
+ inference_models/models/clip/clip_pytorch.py,sha256=yA63g0X_mXUzcKA11gV9XsxvQEocFuK3qsaOgGRy0gg,3837
39
+ inference_models/models/clip/preprocessing.py,sha256=xx0rSGM7ONJud3BCZkkwjGI6uduW_wqLz5qcpzQgkuw,6195
40
+ inference_models/models/common/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
41
+ inference_models/models/common/cuda.py,sha256=nkbRiIIE6Gv077MPCyHHlziUf1xGzVi7fjHjp83JCnY,738
42
+ inference_models/models/common/model_packages.py,sha256=lbtc066teiK0u3mxvmysF5_iZezOVc2VW_HZOZTJlKM,1125
43
+ inference_models/models/common/onnx.py,sha256=VwvdooOMWYN6kvb8j0ZAu3vCxyndeoFO8GStcWUMOto,16022
44
+ inference_models/models/common/torch.py,sha256=rTHWPUqcZ-naM7MRpnPgLMgw-bE0jjpiG_tF81PEJCc,662
45
+ inference_models/models/common/trt.py,sha256=Xeq0wPx5kJli7l9ZLuRkzw_rw7FycHEoXNDts9RkFUs,10086
46
+ inference_models/models/common/roboflow/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
47
+ inference_models/models/common/roboflow/model_packages.py,sha256=a5SqBFr3YS4ML-yF-YRWNJOdEbptTfRi9HeN61O2JEE,12792
48
+ inference_models/models/common/roboflow/post_processing.py,sha256=_1bCiEyWJnBIBWLGaaBvOYJ8a61sDf72injFv3lzhoA,15417
49
+ inference_models/models/common/roboflow/pre_processing.py,sha256=JZzc38I40Iy3TNjRvs-UNIIYd9GeKxQ7WyiamFWsQLc,54073
50
+ inference_models/models/deep_lab_v3_plus/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
51
+ inference_models/models/deep_lab_v3_plus/deep_lab_v3_plus_segmentation_onnx.py,sha256=TsClex1x-tE1TVVKELEuJ_7W3v7PdfWm7n5_8LmOtrs,11948
52
+ inference_models/models/deep_lab_v3_plus/deep_lab_v3_plus_segmentation_torch.py,sha256=zHrJgzvut3_kLRpz69PwPtCwjGq2b01TOLcZY7orsbU,11072
53
+ inference_models/models/deep_lab_v3_plus/deep_lab_v3_plus_segmentation_trt.py,sha256=ZCaQaZHsT3dtf1OpTl90RR224xSayOs1bXHYELBBt2Q,12763
54
+ inference_models/models/depth_anything_v2/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
55
+ inference_models/models/depth_anything_v2/depth_anything_v2_hf.py,sha256=GEedyZAdTj7EiSO6YsRNjWXztqdOCwyfRne02WRxDZg,2553
56
+ inference_models/models/dinov3/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
57
+ inference_models/models/dinov3/dinov3_classification_onnx.py,sha256=MkR2TMr6Cp_SH2J6B8vaDojVKGC8_sijL7na2sO6ys8,13120
58
+ inference_models/models/dinov3/dinov3_classification_torch.py,sha256=pngI1Mmw42YVaB5T_xUk658eg-Pbj3wNiWLrNE8C7s0,11363
59
+ inference_models/models/doctr/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
60
+ inference_models/models/doctr/doctr_torch.py,sha256=pOnXtLrn4sXTfPZN1OjWiMYSjJR4HpqTVAUZZrTrR6c,11614
61
+ inference_models/models/easy_ocr/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
62
+ inference_models/models/easy_ocr/easy_ocr_torch.py,sha256=zD54dnPNEEHyxyIzY3VCduTwSinkZ3aBASzF1c3mPhA,8805
63
+ inference_models/models/florence2/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
64
+ inference_models/models/florence2/florence2_hf.py,sha256=V3rzZPdom-sDjSsTlwf_yjy1QoMcuqvyl5TcMmn3pqU,34330
65
+ inference_models/models/grounding_dino/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
66
+ inference_models/models/grounding_dino/grounding_dino_torch.py,sha256=O4aqxe0G774_OpP1u-cp3bwj9T8zPM6AeBin1_bDaZI,8896
67
+ inference_models/models/l2cs/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
68
+ inference_models/models/l2cs/l2cs_onnx.py,sha256=DC3emDPXZYwxQb9AwW7IYN6vyaT5CVYS4_JJDrHg-BI,8440
69
+ inference_models/models/mediapipe_face_detection/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
70
+ inference_models/models/mediapipe_face_detection/face_detection.py,sha256=JoDS6EUWEc-IE_wSDgEEn4jJd5CPekJDgHLBCz4JJtk,8920
71
+ inference_models/models/moondream2/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
72
+ inference_models/models/moondream2/moondream2_hf.py,sha256=1JdpEux3PajBZ7Up9ReHp5Nj3MMwGVE1S_QF1QAz4PI,9447
73
+ inference_models/models/owlv2/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
74
+ inference_models/models/owlv2/cache.py,sha256=KZPpkUGgVwuZk4XOsDqGB9shM6naf2j4GZApsJE5i6M,5795
75
+ inference_models/models/owlv2/entities.py,sha256=F5iD2k7naSrC04Ub7RUWKT2Dg-LyMbW2Rhmk1EWNUhc,3683
76
+ inference_models/models/owlv2/owlv2_hf.py,sha256=WcqS9Xsmh3OQvxgKejRgDgIR0s6ymU0sCfS2k4lZPlc,28928
77
+ inference_models/models/owlv2/reference_dataset.py,sha256=TkekaI94P9BVWv3IZNBmOSKM1ihZyyp65ShTvThBTok,10675
78
+ inference_models/models/paligemma/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
79
+ inference_models/models/paligemma/paligemma_hf.py,sha256=cs_INy7qaYQAPsM1zcug0CZVf4328cdsWsl9dzHwIfk,7498
80
+ inference_models/models/perception_encoder/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
81
+ inference_models/models/perception_encoder/perception_encoder_pytorch.py,sha256=RBcMCQQnqsSV8YiEKqdQb1woTaRKAIQahcWIfhvMhxQ,7535
82
+ inference_models/models/perception_encoder/vision_encoder/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
83
+ inference_models/models/perception_encoder/vision_encoder/config.py,sha256=VLf_sLKCsGz1wIJAaEsRP6I6jMYHRgfrikH92QFfDSY,3500
84
+ inference_models/models/perception_encoder/vision_encoder/pe.py,sha256=6VL7tJMTt0YxKt20g8s2yq7BKmA8YsKUD7mhhqP9WEM,23709
85
+ inference_models/models/perception_encoder/vision_encoder/rope.py,sha256=gNOzAZzQ-3xqmJQgpnpPBNfWi-riqOe3BsSowJnvtLA,10502
86
+ inference_models/models/perception_encoder/vision_encoder/tokenizer.py,sha256=doV4hmh4bk1VnbvFlK4U5sY06Ky9Bx89Uti89RRoIj0,11542
87
+ inference_models/models/perception_encoder/vision_encoder/transforms.py,sha256=z5uc86siRWYSfXV5nL1B67p4hGUj35xSocNPq2LpTNI,922
88
+ inference_models/models/qwen25vl/__init__.py,sha256=ivA02JoS6FqnY-bi-cst9H7j7JmGtQPZQS3EIkF5DI8,58
89
+ inference_models/models/qwen25vl/qwen25vl_hf.py,sha256=mJp8_X82eU8e2dAJERuDuMHhbegN21HYDjQ_luRFcsI,10033
90
+ inference_models/models/resnet/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
91
+ inference_models/models/resnet/resnet_classification_onnx.py,sha256=Mv2aqQlRhKzMCJxrPDJ7u6hubONvJvhPfUQ3nt_xkqI,12679
92
+ inference_models/models/resnet/resnet_classification_torch.py,sha256=sxfhdyhidElNni-l-kSzNBqI4SMuMt6XrQieJsrUvEc,10795
93
+ inference_models/models/resnet/resnet_classification_trt.py,sha256=WvS2HtpDjNqTLmutcD5bXpYgWvOexFAcW5wRTDxeuuo,13528
94
+ inference_models/models/rfdetr/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
95
+ inference_models/models/rfdetr/backbone_builder.py,sha256=PqhljbZB2b88oJ0aSf4jJrEUcB2iHy1OEWMricY1e98,2866
96
+ inference_models/models/rfdetr/class_remapping.py,sha256=JXTBF7uCCmFHnjO4ffmagERk6JT-JO5HrHu-2Z0QKKM,1393
97
+ inference_models/models/rfdetr/common.py,sha256=oNk6T5rCrnvpW3QhBqzWo1zj7IoWSXeEvwI5WhaNGSA,4679
98
+ inference_models/models/rfdetr/default_labels.py,sha256=BAFZJwZLvvN8Ui4Nxc2ScbEhdiClo3AwFwV6Zd0vR8s,1849
99
+ inference_models/models/rfdetr/dinov2_with_windowed_attn.py,sha256=2wYXXVaqNK45dCowUGsHD63VXcRtZdPHGjv_V_yUd4w,54220
100
+ inference_models/models/rfdetr/misc.py,sha256=EhzPmWNA5yDJAMQN0E05i6e5BI6Scow53eVrpNc1Pus,666
101
+ inference_models/models/rfdetr/ms_deform_attn.py,sha256=IkrRLMel05bz0SLQGGK-vNJuwkBtiHeEp12Tc3Xqq8w,7304
102
+ inference_models/models/rfdetr/ms_deform_attn_func.py,sha256=j5GzCZjgG8bWPTBKpcNVG2SC9yUJ0HFPWAogKPck5HQ,2706
103
+ inference_models/models/rfdetr/position_encoding.py,sha256=aZAlbgh7QYjvd9c_5FTUv_FSie3UkHY22cnLIcs6od0,6104
104
+ inference_models/models/rfdetr/post_processor.py,sha256=Lq0jfmORWTBRLfPI9Hf3bQbigafivGuaLy9nosmKdxI,3071
105
+ inference_models/models/rfdetr/projector.py,sha256=JZHQxRTj_8lt0Bo2M2aq4bSKtOk9x3UyuLVgA8LYXWA,12684
106
+ inference_models/models/rfdetr/rfdetr_backbone_pytorch.py,sha256=e5K9rhKsGMAhU1UxfLib6ivgsxCHDWc7TrOVXeJ7B0A,14209
107
+ inference_models/models/rfdetr/rfdetr_base_pytorch.py,sha256=eFMmwLYhaZ8yNCoMhCkngTaU3cxiEdX_a-ueHt7bapA,28971
108
+ inference_models/models/rfdetr/rfdetr_instance_segmentation_onnx.py,sha256=3lFZLn6HdOtEHo1op_1UN9tvsA-f7T4xL3KPX3RarLg,8024
109
+ inference_models/models/rfdetr/rfdetr_instance_segmentation_pytorch.py,sha256=YUfTRAo_2GOVOpbu6Csy8NWKraDvZXK6L-bfuYxB22s,14926
110
+ inference_models/models/rfdetr/rfdetr_instance_segmentation_trt.py,sha256=qieNBjkhmbbKQyvUGTdtuP93Hn3rDFqYx9PoWPMkE6k,8398
111
+ inference_models/models/rfdetr/rfdetr_object_detection_onnx.py,sha256=MA7-ww9h9YRtx7WlPfgElZhQmnPpZL8TOiNfUogmUy8,9765
112
+ inference_models/models/rfdetr/rfdetr_object_detection_pytorch.py,sha256=UMYQvXayNZnyWVNIN3V9mcXbooqj5xsxPHb3S94pN6A,19329
113
+ inference_models/models/rfdetr/rfdetr_object_detection_trt.py,sha256=XZGRooT8qywjTVt-6IUX5oogHezu9QfD_rcGeyzCeHQ,10423
114
+ inference_models/models/rfdetr/segmentation_head.py,sha256=qg0m09xJVQ_DP5lYx_IkYbuC3UsLcv915cMQx65Alio,10214
115
+ inference_models/models/rfdetr/transformer.py,sha256=aYbE9x6QBA5T56pAE9lQ_lgMreG3INdAT5LaywOto9c,27502
116
+ inference_models/models/roboflow_instant/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
117
+ inference_models/models/roboflow_instant/roboflow_instant_hf.py,sha256=UN5_wIhCZZjBmPKzcP0MDAQbEizcYSnufAga2Uru3cM,5309
118
+ inference_models/models/sam/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
119
+ inference_models/models/sam/cache.py,sha256=ZtQnhLBOeyVnaknMLErueovcFheaM1cbXtg5pt2h_i8,4742
120
+ inference_models/models/sam/entities.py,sha256=72WR3mVkxGJeRoZbcObuS5Ay6Mw8cqMtqqfyKm6DLfQ,584
121
+ inference_models/models/sam/sam_torch.py,sha256=gOPJT4Fi-kqrhFZrlblm1MZ17rg2MD9cIWqJFplK9Xo,29397
122
+ inference_models/models/sam2/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
123
+ inference_models/models/sam2/cache.py,sha256=2tNkn_WGaZnhmc24cMZd3zmGP2nYFzW_IVEh0A-61Gc,5609
124
+ inference_models/models/sam2/entities.py,sha256=33TZpSv1bv5MWkZu77ksgp0RGO-fz93_J5NqtiweAMQ,1154
125
+ inference_models/models/sam2/sam2_torch.py,sha256=8yaXHnYRklOrpiP6f8WuJzFnaoNWlt0n6R0kM3Kb4JM,38787
126
+ inference_models/models/sam2_rt/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
127
+ inference_models/models/sam2_rt/sam2_pytorch.py,sha256=-hV_osk_RCzRj1delAYaTFneZ-nUP3lGNwtTmX5Uzm8,4482
128
+ inference_models/models/smolvlm/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
129
+ inference_models/models/smolvlm/smolvlm_hf.py,sha256=nwAFziRoxlK-pwnF6vIwB331IIW_H57v8t-Ai5h6fH0,8903
130
+ inference_models/models/trocr/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
131
+ inference_models/models/trocr/trocr_hf.py,sha256=rFdTySKEH5pyBneQrGZA4_4Ix9-SZ1zp8YZ7ZFAmeUQ,1786
132
+ inference_models/models/vit/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
133
+ inference_models/models/vit/vit_classification_huggingface.py,sha256=Lq5Z9X4i4ghpxF0QzLQduJ5n2r6OulOCLyJg0x4kLG4,11027
134
+ inference_models/models/vit/vit_classification_onnx.py,sha256=ml05Mq4YpC5Ibca5AX_y54c43Q87UU5lKev82BpHYT4,12473
135
+ inference_models/models/vit/vit_classification_trt.py,sha256=1R8jJb6tMFejCEJje2pzoFGqoj4FYmQwJSu1OiTvLBY,13325
136
+ inference_models/models/yolact/__init__.py,sha256=wMN3BTunf5CFJH_sV_LIgMUufYfiFwVZUjbaBbdlq1M,33
137
+ inference_models/models/yolact/yolact_instance_segmentation_onnx.py,sha256=SK3-uC4aJaZtfE89KCD1E9hmySIx50hIT8py0pu_YnM,13272
138
+ inference_models/models/yolact/yolact_instance_segmentation_trt.py,sha256=APh0QS4WELptxpaHBiqF6PYZhouks7gTZcoLxJEeE2s,13770
139
+ inference_models/models/yolo_world/__init__.py,sha256=sQfPkM6WqydxTa9ORxqCaSjsVUjCWSIh8_Fp6VcWudM,73
140
+ inference_models/models/yolonas/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
141
+ inference_models/models/yolonas/nms.py,sha256=kYLONdjr8vp9yjjpst8jiD_16qMM2l3ArXd35bL_cQo,1443
142
+ inference_models/models/yolonas/yolonas_object_detection_onnx.py,sha256=-OqZQs8pvke65teh09H6AsjQ4z8kxLa1kKK98x45ZJQ,8047
143
+ inference_models/models/yolonas/yolonas_object_detection_trt.py,sha256=rHz5Z9QSkUnyDqKohDZnArknytybDBjKsCYudFBPH2o,8632
144
+ inference_models/models/yolov10/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
145
+ inference_models/models/yolov10/yolov10_object_detection_onnx.py,sha256=dj4er3d2p7vP5AYaTNEMi0Dy9qd2AQvukN1lf8BlOLk,7272
146
+ inference_models/models/yolov10/yolov10_object_detection_trt.py,sha256=J71uE4xCEUtoFs8fXwDkIfVFsdHTckR5gA8zTE76Pxs,7893
147
+ inference_models/models/yolov11/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
148
+ inference_models/models/yolov11/yolov11_onnx.py,sha256=9wYERJDCjxa6BxotjmOCKN9Nf2tEVnCcGjCmJ_cl-bI,783
149
+ inference_models/models/yolov11/yolov11_torch_script.py,sha256=i1CFesljr4t3OtWVGk_WzPEJvbvDxnQgL0EFj4vH8Ww,699
150
+ inference_models/models/yolov11/yolov11_trt.py,sha256=vhg7MzfEQ7rpGUVvQI_MmS25GwR8UjwVKgRHmLTOu2M,588
151
+ inference_models/models/yolov12/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
152
+ inference_models/models/yolov12/yolov12_onnx.py,sha256=sSW8U7z_hHq2Rq_vdG2qbrPikNEaAm29M6rs4X_CSeo,188
153
+ inference_models/models/yolov12/yolov12_torch_script.py,sha256=pH7dFi1JgFCt8oIBe0gmeBvsdMWDU2R1z4icWVBkToI,217
154
+ inference_models/models/yolov12/yolov12_trt.py,sha256=vLhn2ZJfAmgZW7UiowmZosUMo-XhIWrZeXhWbT0Rx-c,184
155
+ inference_models/models/yolov5/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
156
+ inference_models/models/yolov5/nms.py,sha256=3uu8JpBBZTUkaSfcVziYvQ3QhXTrfDmOCs2adkR8D6U,3458
157
+ inference_models/models/yolov5/yolov5_instance_segmentation_onnx.py,sha256=eVKQFVUDz8B1bWdkBYG02yzX0ya6-WDKQS6l5Cr5Ico,8873
158
+ inference_models/models/yolov5/yolov5_instance_segmentation_trt.py,sha256=E285z4QtO1I6Zud18w-X-9ts_XMIUU8dyxEELqrqlGA,9724
159
+ inference_models/models/yolov5/yolov5_object_detection_onnx.py,sha256=TTwAZYVWBz8VXug0jHKxWOKB8vMQrvhnb0opXDYg67M,7452
160
+ inference_models/models/yolov5/yolov5_object_detection_trt.py,sha256=bgp4ubTxwoi14QXhXuNsILdHKFsZ6fikjg9gdP76ViA,8007
161
+ inference_models/models/yolov7/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
162
+ inference_models/models/yolov7/yolov7_instance_segmentation_onnx.py,sha256=GhZx8Dx_QEVFU9A1prydFP2TRPXobGIE72dt6v_B1XA,8874
163
+ inference_models/models/yolov7/yolov7_instance_segmentation_trt.py,sha256=YXOG78aZ_3koNKu_kyUcqifXaKyud15a-P6-PRI4Lvw,9478
164
+ inference_models/models/yolov8/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
165
+ inference_models/models/yolov8/yolov8_classification_onnx.py,sha256=fMi-mkdzsUi9eWfcdn5_iEQ-D5LqxKAJaHzC8tcj0wo,7111
166
+ inference_models/models/yolov8/yolov8_instance_segmentation_onnx.py,sha256=CL2tNH09TjO-PCIONBbALt8fy113HLNz2f0oY_x3dS4,9425
167
+ inference_models/models/yolov8/yolov8_instance_segmentation_torch_script.py,sha256=q6qZdVX7TimnyqzBDlk0uFykWZKji16S-gyvlCOmhrs,7701
168
+ inference_models/models/yolov8/yolov8_instance_segmentation_trt.py,sha256=i7tOfSdKVGANCKXJx0_0E03yVzrtK2CrGuOpKbyx_ks,10178
169
+ inference_models/models/yolov8/yolov8_key_points_detection_onnx.py,sha256=IKQANr9fxIEwnQ7K7QNi4NYKtiC0yGiU_aduBUGkXLs,10612
170
+ inference_models/models/yolov8/yolov8_key_points_detection_torch_script.py,sha256=986Bx2uSG9eBEyaBv8G50OOZoT3trrls7pzguttFTgo,8552
171
+ inference_models/models/yolov8/yolov8_key_points_detection_trt.py,sha256=xl0ha3onxR8BqtnLTuUEDdaJIsJpyxBL8FWMYT0TXis,11139
172
+ inference_models/models/yolov8/yolov8_object_detection_onnx.py,sha256=oJR1OmiWUemO5RwaoePO-IwXk1aQsxWV0Bhhd0MvoeM,8251
173
+ inference_models/models/yolov8/yolov8_object_detection_torch_script.py,sha256=M3R5fcrTRa0xkXCna9OkiOqthReANi7Qm2TuI-5Gs4o,6104
174
+ inference_models/models/yolov8/yolov8_object_detection_trt.py,sha256=pt0dHffNR6KVFXgy05EKwcMXP1FJfMqoChiYkNJ-gTk,8478
175
+ inference_models/models/yolov9/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
176
+ inference_models/models/yolov9/yolov9_onnx.py,sha256=zV7LK3CtSVR1R9zKJjCdVUvmCbDNg66PDI1mgorlk74,187
177
+ inference_models/models/yolov9/yolov9_torch_script.py,sha256=YmOE3ioUrLvU_SyPQxfIrErNrbBUxDUcyHHPx6IlW7c,216
178
+ inference_models/models/yolov9/yolov9_trt.py,sha256=l1PUidw005b0hUF2VAd45GV6ICSwxK48Txt_P_InVIg,183
179
+ inference_models/runtime_introspection/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
180
+ inference_models/runtime_introspection/core.py,sha256=9_XaCekjX1OROlSNsZ4q68iImhn6pbxqtJE0a_4GDk4,13230
181
+ inference_models/utils/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
182
+ inference_models/utils/download.py,sha256=YoVvZBIJAiJu2OqBJfDiYBcih8M4tAGUD9QFB5w7Dl4,21950
183
+ inference_models/utils/environment.py,sha256=koaOp-b6ovv_mCLVhO7wzIX26Pg9wi0oBj6vWoOhP4I,853
184
+ inference_models/utils/file_system.py,sha256=Z5s2AmNeTVvzM6OHy8fJPsboqQFttbzTMpUV7Ibg6SU,1405
185
+ inference_models/utils/hashing.py,sha256=Gb0ISj6cc32aZ94oI9BwtIEw6AYOSJaL2YiGycFQVRg,194
186
+ inference_models/utils/imports.py,sha256=ZqsVmdVVdfl3V1Ke7HRnXv0CIUE-6J0RlGCef9EfdcA,1514
187
+ inference_models/utils/onnx_introspection.py,sha256=sxQCdDWXtPokUgMhK4mN_Hxiv4ea85aGAnPnt63Z6q4,460
188
+ inference_models/weights_providers/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
189
+ inference_models/weights_providers/core.py,sha256=ZCm7R-svrgaiOZ3AXpGA5OsYcDwhtwGRN2la5mpG_GA,715
190
+ inference_models/weights_providers/entities.py,sha256=hI-tQ7BkdabnBfMko2wYo76jRGWXJUWlSTwq2OQ4RDM,5834
191
+ inference_models/weights_providers/roboflow.py,sha256=6tXzqolMnQLIz5vR_eXQsZ1DmZ9UMPlimddZLQFA_8Y,25358
192
+ inference_models-0.18.3.dist-info/METADATA,sha256=QCLAbiUACa97jaTjKGGDVTq-Maa31PS9GtIqjlxldgU,20940
193
+ inference_models-0.18.3.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
194
+ inference_models-0.18.3.dist-info/top_level.txt,sha256=128IRjtd3k9I4c7i-eErAE2oOlOIF6-6iRW7tHm7VxE,17
195
+ inference_models-0.18.3.dist-info/RECORD,,
@@ -0,0 +1,5 @@
1
+ Wheel-Version: 1.0
2
+ Generator: setuptools (80.9.0)
3
+ Root-Is-Purelib: true
4
+ Tag: py3-none-any
5
+
@@ -0,0 +1 @@
1
+ inference_models