inference-models 0.18.3__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- inference_models/__init__.py +36 -0
- inference_models/configuration.py +72 -0
- inference_models/constants.py +2 -0
- inference_models/entities.py +5 -0
- inference_models/errors.py +137 -0
- inference_models/logger.py +52 -0
- inference_models/model_pipelines/__init__.py +0 -0
- inference_models/model_pipelines/auto_loaders/__init__.py +0 -0
- inference_models/model_pipelines/auto_loaders/core.py +120 -0
- inference_models/model_pipelines/auto_loaders/pipelines_registry.py +36 -0
- inference_models/model_pipelines/face_and_gaze_detection/__init__.py +0 -0
- inference_models/model_pipelines/face_and_gaze_detection/mediapipe_l2cs.py +200 -0
- inference_models/models/__init__.py +0 -0
- inference_models/models/auto_loaders/__init__.py +0 -0
- inference_models/models/auto_loaders/access_manager.py +168 -0
- inference_models/models/auto_loaders/auto_negotiation.py +1329 -0
- inference_models/models/auto_loaders/auto_resolution_cache.py +129 -0
- inference_models/models/auto_loaders/constants.py +7 -0
- inference_models/models/auto_loaders/core.py +1341 -0
- inference_models/models/auto_loaders/dependency_models.py +52 -0
- inference_models/models/auto_loaders/entities.py +57 -0
- inference_models/models/auto_loaders/models_registry.py +497 -0
- inference_models/models/auto_loaders/presentation_utils.py +333 -0
- inference_models/models/auto_loaders/ranking.py +413 -0
- inference_models/models/auto_loaders/utils.py +31 -0
- inference_models/models/base/__init__.py +0 -0
- inference_models/models/base/classification.py +123 -0
- inference_models/models/base/depth_estimation.py +62 -0
- inference_models/models/base/documents_parsing.py +111 -0
- inference_models/models/base/embeddings.py +66 -0
- inference_models/models/base/instance_segmentation.py +87 -0
- inference_models/models/base/keypoints_detection.py +93 -0
- inference_models/models/base/object_detection.py +143 -0
- inference_models/models/base/semantic_segmentation.py +74 -0
- inference_models/models/base/types.py +5 -0
- inference_models/models/clip/__init__.py +0 -0
- inference_models/models/clip/clip_onnx.py +148 -0
- inference_models/models/clip/clip_pytorch.py +104 -0
- inference_models/models/clip/preprocessing.py +162 -0
- inference_models/models/common/__init__.py +0 -0
- inference_models/models/common/cuda.py +30 -0
- inference_models/models/common/model_packages.py +25 -0
- inference_models/models/common/onnx.py +379 -0
- inference_models/models/common/roboflow/__init__.py +0 -0
- inference_models/models/common/roboflow/model_packages.py +361 -0
- inference_models/models/common/roboflow/post_processing.py +436 -0
- inference_models/models/common/roboflow/pre_processing.py +1332 -0
- inference_models/models/common/torch.py +20 -0
- inference_models/models/common/trt.py +266 -0
- inference_models/models/deep_lab_v3_plus/__init__.py +0 -0
- inference_models/models/deep_lab_v3_plus/deep_lab_v3_plus_segmentation_onnx.py +282 -0
- inference_models/models/deep_lab_v3_plus/deep_lab_v3_plus_segmentation_torch.py +264 -0
- inference_models/models/deep_lab_v3_plus/deep_lab_v3_plus_segmentation_trt.py +313 -0
- inference_models/models/depth_anything_v2/__init__.py +0 -0
- inference_models/models/depth_anything_v2/depth_anything_v2_hf.py +77 -0
- inference_models/models/dinov3/__init__.py +0 -0
- inference_models/models/dinov3/dinov3_classification_onnx.py +348 -0
- inference_models/models/dinov3/dinov3_classification_torch.py +323 -0
- inference_models/models/doctr/__init__.py +0 -0
- inference_models/models/doctr/doctr_torch.py +304 -0
- inference_models/models/easy_ocr/__init__.py +0 -0
- inference_models/models/easy_ocr/easy_ocr_torch.py +222 -0
- inference_models/models/florence2/__init__.py +0 -0
- inference_models/models/florence2/florence2_hf.py +897 -0
- inference_models/models/grounding_dino/__init__.py +0 -0
- inference_models/models/grounding_dino/grounding_dino_torch.py +227 -0
- inference_models/models/l2cs/__init__.py +0 -0
- inference_models/models/l2cs/l2cs_onnx.py +216 -0
- inference_models/models/mediapipe_face_detection/__init__.py +0 -0
- inference_models/models/mediapipe_face_detection/face_detection.py +203 -0
- inference_models/models/moondream2/__init__.py +0 -0
- inference_models/models/moondream2/moondream2_hf.py +281 -0
- inference_models/models/owlv2/__init__.py +0 -0
- inference_models/models/owlv2/cache.py +182 -0
- inference_models/models/owlv2/entities.py +112 -0
- inference_models/models/owlv2/owlv2_hf.py +695 -0
- inference_models/models/owlv2/reference_dataset.py +291 -0
- inference_models/models/paligemma/__init__.py +0 -0
- inference_models/models/paligemma/paligemma_hf.py +209 -0
- inference_models/models/perception_encoder/__init__.py +0 -0
- inference_models/models/perception_encoder/perception_encoder_pytorch.py +197 -0
- inference_models/models/perception_encoder/vision_encoder/__init__.py +0 -0
- inference_models/models/perception_encoder/vision_encoder/config.py +160 -0
- inference_models/models/perception_encoder/vision_encoder/pe.py +742 -0
- inference_models/models/perception_encoder/vision_encoder/rope.py +344 -0
- inference_models/models/perception_encoder/vision_encoder/tokenizer.py +342 -0
- inference_models/models/perception_encoder/vision_encoder/transforms.py +33 -0
- inference_models/models/qwen25vl/__init__.py +1 -0
- inference_models/models/qwen25vl/qwen25vl_hf.py +285 -0
- inference_models/models/resnet/__init__.py +0 -0
- inference_models/models/resnet/resnet_classification_onnx.py +330 -0
- inference_models/models/resnet/resnet_classification_torch.py +305 -0
- inference_models/models/resnet/resnet_classification_trt.py +369 -0
- inference_models/models/rfdetr/__init__.py +0 -0
- inference_models/models/rfdetr/backbone_builder.py +101 -0
- inference_models/models/rfdetr/class_remapping.py +41 -0
- inference_models/models/rfdetr/common.py +115 -0
- inference_models/models/rfdetr/default_labels.py +108 -0
- inference_models/models/rfdetr/dinov2_with_windowed_attn.py +1330 -0
- inference_models/models/rfdetr/misc.py +26 -0
- inference_models/models/rfdetr/ms_deform_attn.py +180 -0
- inference_models/models/rfdetr/ms_deform_attn_func.py +60 -0
- inference_models/models/rfdetr/position_encoding.py +166 -0
- inference_models/models/rfdetr/post_processor.py +83 -0
- inference_models/models/rfdetr/projector.py +373 -0
- inference_models/models/rfdetr/rfdetr_backbone_pytorch.py +394 -0
- inference_models/models/rfdetr/rfdetr_base_pytorch.py +807 -0
- inference_models/models/rfdetr/rfdetr_instance_segmentation_onnx.py +206 -0
- inference_models/models/rfdetr/rfdetr_instance_segmentation_pytorch.py +373 -0
- inference_models/models/rfdetr/rfdetr_instance_segmentation_trt.py +227 -0
- inference_models/models/rfdetr/rfdetr_object_detection_onnx.py +244 -0
- inference_models/models/rfdetr/rfdetr_object_detection_pytorch.py +470 -0
- inference_models/models/rfdetr/rfdetr_object_detection_trt.py +270 -0
- inference_models/models/rfdetr/segmentation_head.py +273 -0
- inference_models/models/rfdetr/transformer.py +767 -0
- inference_models/models/roboflow_instant/__init__.py +0 -0
- inference_models/models/roboflow_instant/roboflow_instant_hf.py +141 -0
- inference_models/models/sam/__init__.py +0 -0
- inference_models/models/sam/cache.py +147 -0
- inference_models/models/sam/entities.py +25 -0
- inference_models/models/sam/sam_torch.py +675 -0
- inference_models/models/sam2/__init__.py +0 -0
- inference_models/models/sam2/cache.py +162 -0
- inference_models/models/sam2/entities.py +43 -0
- inference_models/models/sam2/sam2_torch.py +905 -0
- inference_models/models/sam2_rt/__init__.py +0 -0
- inference_models/models/sam2_rt/sam2_pytorch.py +119 -0
- inference_models/models/smolvlm/__init__.py +0 -0
- inference_models/models/smolvlm/smolvlm_hf.py +245 -0
- inference_models/models/trocr/__init__.py +0 -0
- inference_models/models/trocr/trocr_hf.py +53 -0
- inference_models/models/vit/__init__.py +0 -0
- inference_models/models/vit/vit_classification_huggingface.py +319 -0
- inference_models/models/vit/vit_classification_onnx.py +326 -0
- inference_models/models/vit/vit_classification_trt.py +365 -0
- inference_models/models/yolact/__init__.py +1 -0
- inference_models/models/yolact/yolact_instance_segmentation_onnx.py +336 -0
- inference_models/models/yolact/yolact_instance_segmentation_trt.py +361 -0
- inference_models/models/yolo_world/__init__.py +1 -0
- inference_models/models/yolonas/__init__.py +0 -0
- inference_models/models/yolonas/nms.py +44 -0
- inference_models/models/yolonas/yolonas_object_detection_onnx.py +204 -0
- inference_models/models/yolonas/yolonas_object_detection_trt.py +230 -0
- inference_models/models/yolov10/__init__.py +0 -0
- inference_models/models/yolov10/yolov10_object_detection_onnx.py +187 -0
- inference_models/models/yolov10/yolov10_object_detection_trt.py +215 -0
- inference_models/models/yolov11/__init__.py +0 -0
- inference_models/models/yolov11/yolov11_onnx.py +28 -0
- inference_models/models/yolov11/yolov11_torch_script.py +25 -0
- inference_models/models/yolov11/yolov11_trt.py +21 -0
- inference_models/models/yolov12/__init__.py +0 -0
- inference_models/models/yolov12/yolov12_onnx.py +7 -0
- inference_models/models/yolov12/yolov12_torch_script.py +7 -0
- inference_models/models/yolov12/yolov12_trt.py +7 -0
- inference_models/models/yolov5/__init__.py +0 -0
- inference_models/models/yolov5/nms.py +99 -0
- inference_models/models/yolov5/yolov5_instance_segmentation_onnx.py +225 -0
- inference_models/models/yolov5/yolov5_instance_segmentation_trt.py +255 -0
- inference_models/models/yolov5/yolov5_object_detection_onnx.py +192 -0
- inference_models/models/yolov5/yolov5_object_detection_trt.py +218 -0
- inference_models/models/yolov7/__init__.py +0 -0
- inference_models/models/yolov7/yolov7_instance_segmentation_onnx.py +226 -0
- inference_models/models/yolov7/yolov7_instance_segmentation_trt.py +253 -0
- inference_models/models/yolov8/__init__.py +0 -0
- inference_models/models/yolov8/yolov8_classification_onnx.py +181 -0
- inference_models/models/yolov8/yolov8_instance_segmentation_onnx.py +239 -0
- inference_models/models/yolov8/yolov8_instance_segmentation_torch_script.py +201 -0
- inference_models/models/yolov8/yolov8_instance_segmentation_trt.py +268 -0
- inference_models/models/yolov8/yolov8_key_points_detection_onnx.py +263 -0
- inference_models/models/yolov8/yolov8_key_points_detection_torch_script.py +218 -0
- inference_models/models/yolov8/yolov8_key_points_detection_trt.py +287 -0
- inference_models/models/yolov8/yolov8_object_detection_onnx.py +213 -0
- inference_models/models/yolov8/yolov8_object_detection_torch_script.py +166 -0
- inference_models/models/yolov8/yolov8_object_detection_trt.py +231 -0
- inference_models/models/yolov9/__init__.py +0 -0
- inference_models/models/yolov9/yolov9_onnx.py +7 -0
- inference_models/models/yolov9/yolov9_torch_script.py +7 -0
- inference_models/models/yolov9/yolov9_trt.py +7 -0
- inference_models/runtime_introspection/__init__.py +0 -0
- inference_models/runtime_introspection/core.py +410 -0
- inference_models/utils/__init__.py +0 -0
- inference_models/utils/download.py +608 -0
- inference_models/utils/environment.py +28 -0
- inference_models/utils/file_system.py +51 -0
- inference_models/utils/hashing.py +7 -0
- inference_models/utils/imports.py +48 -0
- inference_models/utils/onnx_introspection.py +17 -0
- inference_models/weights_providers/__init__.py +0 -0
- inference_models/weights_providers/core.py +20 -0
- inference_models/weights_providers/entities.py +159 -0
- inference_models/weights_providers/roboflow.py +601 -0
- inference_models-0.18.3.dist-info/METADATA +466 -0
- inference_models-0.18.3.dist-info/RECORD +195 -0
- inference_models-0.18.3.dist-info/WHEEL +5 -0
- inference_models-0.18.3.dist-info/top_level.txt +1 -0
|
@@ -0,0 +1,195 @@
|
|
|
1
|
+
inference_models/__init__.py,sha256=VQl38l8GyUdRb0H0tB8eYrv_8831HmNLsNtOLCY30Uc,1339
|
|
2
|
+
inference_models/configuration.py,sha256=lYkdnzLAYwVzaGUDnRaNYWKqBVurq52zz4ASfQQWoEw,2630
|
|
3
|
+
inference_models/constants.py,sha256=644eUneU8TamCzWZ_TJLIv2pPeaHYcFGgXSSNYS2yck,79
|
|
4
|
+
inference_models/entities.py,sha256=xhFqW0z-bWLITIjZEPMStJgw7i97z5STMgQm1MRvK4Y,168
|
|
5
|
+
inference_models/errors.py,sha256=mXGurGA4q3PRjwStm05EZtJP4xCBrYoOFiTHvSQfDUk,2469
|
|
6
|
+
inference_models/logger.py,sha256=pJR922npIgD_cjU6ePYrDMx6_QvzArB5VjM-MDLdv8U,1368
|
|
7
|
+
inference_models/model_pipelines/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
8
|
+
inference_models/model_pipelines/auto_loaders/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
9
|
+
inference_models/model_pipelines/auto_loaders/core.py,sha256=8Mnd5wODg_3cqxgMdP_hAEkPecNo9_99-VCtcfhTm-w,5773
|
|
10
|
+
inference_models/model_pipelines/auto_loaders/pipelines_registry.py,sha256=33PIoVd74t9dnKizs3VhIkLgZx-MLfFTKe6LOl1Q9cY,1290
|
|
11
|
+
inference_models/model_pipelines/face_and_gaze_detection/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
12
|
+
inference_models/model_pipelines/face_and_gaze_detection/mediapipe_l2cs.py,sha256=4bEnp0KunSOihQK9Xc4lqnRJVRtkCcCpVV5-ldBhY-E,8128
|
|
13
|
+
inference_models/models/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
14
|
+
inference_models/models/auto_loaders/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
15
|
+
inference_models/models/auto_loaders/access_manager.py,sha256=Se3pCczKCOTzKBIxVmjK8eiifUgp8lipg7FtM_t0RUY,4196
|
|
16
|
+
inference_models/models/auto_loaders/auto_negotiation.py,sha256=zLjs6G_Bnx0K0e3ls6P6KPrRw3oTgRZTm3bTgXraM4U,60006
|
|
17
|
+
inference_models/models/auto_loaders/auto_resolution_cache.py,sha256=wf6pQXeKIMdkiSEyphQSpeS4EHOeUYLS1r_zXn3sRvo,4836
|
|
18
|
+
inference_models/models/auto_loaders/constants.py,sha256=6-9jnthGIIFvWvFlext-RZmLNKDd-TXLMe6AWsLvw40,306
|
|
19
|
+
inference_models/models/auto_loaders/core.py,sha256=FG9QVIlXkSS9fHkZUnwYjMgJrC4Vqdq7aCFRqmTExVU,64046
|
|
20
|
+
inference_models/models/auto_loaders/dependency_models.py,sha256=PEU3SySjqSHPmwNKEbYihvXum_UtOQjGZ40V0M95Apw,2375
|
|
21
|
+
inference_models/models/auto_loaders/entities.py,sha256=GX7MvMK1geWA2UvP9GR8nUMcazB9IL33xEpwV6xW_JQ,1714
|
|
22
|
+
inference_models/models/auto_loaders/models_registry.py,sha256=xZySoxVkY2i2ewmF3ATnd4aRLYZsZroXuxoTxWbvfU4,22171
|
|
23
|
+
inference_models/models/auto_loaders/presentation_utils.py,sha256=8PXs_tBOEiD_8V6qu0L3a00wc-LZVolKhTnu-1EsDxs,13155
|
|
24
|
+
inference_models/models/auto_loaders/ranking.py,sha256=V846HJ33iycZr_47JOcLW6YCwMoSHhCzCKK2Q_EiXao,16157
|
|
25
|
+
inference_models/models/auto_loaders/utils.py,sha256=mzFl2XGQPXSpEUYgy1pKynwvBbkH3Y3zUwJbfbVUsug,1482
|
|
26
|
+
inference_models/models/base/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
27
|
+
inference_models/models/base/classification.py,sha256=Kh3DNNySyHxrnUGBv85a2rOIuqaPcBWfm_zDjgCvF_U,3484
|
|
28
|
+
inference_models/models/base/depth_estimation.py,sha256=3pQqIVx2JHGXujj_iGYMt8hKL7cSscn3uHw1tvUdK-0,1692
|
|
29
|
+
inference_models/models/base/documents_parsing.py,sha256=aMnDWd921MwziU-4G7YwWJ5e4UNMN2wQWTsTca2tekg,3130
|
|
30
|
+
inference_models/models/base/embeddings.py,sha256=mAGfHfx9cyD0Nig0WMkRaxAL0lyp80n4yVIR1RzmXns,1761
|
|
31
|
+
inference_models/models/base/instance_segmentation.py,sha256=gnWUKMFmSpWqrdjoHUMxAMIrltkAbu1p9QX-GyGPEqg,2482
|
|
32
|
+
inference_models/models/base/keypoints_detection.py,sha256=15Zs6ZKFlbZ3KpQZ75q6ndiPI2OJXt0hUYaObbeNrdk,2712
|
|
33
|
+
inference_models/models/base/object_detection.py,sha256=V-K_N80uwxDmvjYqoTKN74oYqLCyBeiUrcKnWCNp-Ps,4009
|
|
34
|
+
inference_models/models/base/semantic_segmentation.py,sha256=dV9ChC-yfHKNdnGUaG72JOF4Q8AJAB8t623_nIC9hko,1974
|
|
35
|
+
inference_models/models/base/types.py,sha256=2J95RiAwja_flSDr1ldZccqwqn0QQW2c60FNWms54Ag,177
|
|
36
|
+
inference_models/models/clip/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
37
|
+
inference_models/models/clip/clip_onnx.py,sha256=W4cXltU70wMIptiZsr7CX4xzipxZkBRv9rJGIucLkZ4,6205
|
|
38
|
+
inference_models/models/clip/clip_pytorch.py,sha256=yA63g0X_mXUzcKA11gV9XsxvQEocFuK3qsaOgGRy0gg,3837
|
|
39
|
+
inference_models/models/clip/preprocessing.py,sha256=xx0rSGM7ONJud3BCZkkwjGI6uduW_wqLz5qcpzQgkuw,6195
|
|
40
|
+
inference_models/models/common/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
41
|
+
inference_models/models/common/cuda.py,sha256=nkbRiIIE6Gv077MPCyHHlziUf1xGzVi7fjHjp83JCnY,738
|
|
42
|
+
inference_models/models/common/model_packages.py,sha256=lbtc066teiK0u3mxvmysF5_iZezOVc2VW_HZOZTJlKM,1125
|
|
43
|
+
inference_models/models/common/onnx.py,sha256=VwvdooOMWYN6kvb8j0ZAu3vCxyndeoFO8GStcWUMOto,16022
|
|
44
|
+
inference_models/models/common/torch.py,sha256=rTHWPUqcZ-naM7MRpnPgLMgw-bE0jjpiG_tF81PEJCc,662
|
|
45
|
+
inference_models/models/common/trt.py,sha256=Xeq0wPx5kJli7l9ZLuRkzw_rw7FycHEoXNDts9RkFUs,10086
|
|
46
|
+
inference_models/models/common/roboflow/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
47
|
+
inference_models/models/common/roboflow/model_packages.py,sha256=a5SqBFr3YS4ML-yF-YRWNJOdEbptTfRi9HeN61O2JEE,12792
|
|
48
|
+
inference_models/models/common/roboflow/post_processing.py,sha256=_1bCiEyWJnBIBWLGaaBvOYJ8a61sDf72injFv3lzhoA,15417
|
|
49
|
+
inference_models/models/common/roboflow/pre_processing.py,sha256=JZzc38I40Iy3TNjRvs-UNIIYd9GeKxQ7WyiamFWsQLc,54073
|
|
50
|
+
inference_models/models/deep_lab_v3_plus/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
51
|
+
inference_models/models/deep_lab_v3_plus/deep_lab_v3_plus_segmentation_onnx.py,sha256=TsClex1x-tE1TVVKELEuJ_7W3v7PdfWm7n5_8LmOtrs,11948
|
|
52
|
+
inference_models/models/deep_lab_v3_plus/deep_lab_v3_plus_segmentation_torch.py,sha256=zHrJgzvut3_kLRpz69PwPtCwjGq2b01TOLcZY7orsbU,11072
|
|
53
|
+
inference_models/models/deep_lab_v3_plus/deep_lab_v3_plus_segmentation_trt.py,sha256=ZCaQaZHsT3dtf1OpTl90RR224xSayOs1bXHYELBBt2Q,12763
|
|
54
|
+
inference_models/models/depth_anything_v2/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
55
|
+
inference_models/models/depth_anything_v2/depth_anything_v2_hf.py,sha256=GEedyZAdTj7EiSO6YsRNjWXztqdOCwyfRne02WRxDZg,2553
|
|
56
|
+
inference_models/models/dinov3/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
57
|
+
inference_models/models/dinov3/dinov3_classification_onnx.py,sha256=MkR2TMr6Cp_SH2J6B8vaDojVKGC8_sijL7na2sO6ys8,13120
|
|
58
|
+
inference_models/models/dinov3/dinov3_classification_torch.py,sha256=pngI1Mmw42YVaB5T_xUk658eg-Pbj3wNiWLrNE8C7s0,11363
|
|
59
|
+
inference_models/models/doctr/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
60
|
+
inference_models/models/doctr/doctr_torch.py,sha256=pOnXtLrn4sXTfPZN1OjWiMYSjJR4HpqTVAUZZrTrR6c,11614
|
|
61
|
+
inference_models/models/easy_ocr/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
62
|
+
inference_models/models/easy_ocr/easy_ocr_torch.py,sha256=zD54dnPNEEHyxyIzY3VCduTwSinkZ3aBASzF1c3mPhA,8805
|
|
63
|
+
inference_models/models/florence2/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
64
|
+
inference_models/models/florence2/florence2_hf.py,sha256=V3rzZPdom-sDjSsTlwf_yjy1QoMcuqvyl5TcMmn3pqU,34330
|
|
65
|
+
inference_models/models/grounding_dino/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
66
|
+
inference_models/models/grounding_dino/grounding_dino_torch.py,sha256=O4aqxe0G774_OpP1u-cp3bwj9T8zPM6AeBin1_bDaZI,8896
|
|
67
|
+
inference_models/models/l2cs/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
68
|
+
inference_models/models/l2cs/l2cs_onnx.py,sha256=DC3emDPXZYwxQb9AwW7IYN6vyaT5CVYS4_JJDrHg-BI,8440
|
|
69
|
+
inference_models/models/mediapipe_face_detection/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
70
|
+
inference_models/models/mediapipe_face_detection/face_detection.py,sha256=JoDS6EUWEc-IE_wSDgEEn4jJd5CPekJDgHLBCz4JJtk,8920
|
|
71
|
+
inference_models/models/moondream2/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
72
|
+
inference_models/models/moondream2/moondream2_hf.py,sha256=1JdpEux3PajBZ7Up9ReHp5Nj3MMwGVE1S_QF1QAz4PI,9447
|
|
73
|
+
inference_models/models/owlv2/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
74
|
+
inference_models/models/owlv2/cache.py,sha256=KZPpkUGgVwuZk4XOsDqGB9shM6naf2j4GZApsJE5i6M,5795
|
|
75
|
+
inference_models/models/owlv2/entities.py,sha256=F5iD2k7naSrC04Ub7RUWKT2Dg-LyMbW2Rhmk1EWNUhc,3683
|
|
76
|
+
inference_models/models/owlv2/owlv2_hf.py,sha256=WcqS9Xsmh3OQvxgKejRgDgIR0s6ymU0sCfS2k4lZPlc,28928
|
|
77
|
+
inference_models/models/owlv2/reference_dataset.py,sha256=TkekaI94P9BVWv3IZNBmOSKM1ihZyyp65ShTvThBTok,10675
|
|
78
|
+
inference_models/models/paligemma/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
79
|
+
inference_models/models/paligemma/paligemma_hf.py,sha256=cs_INy7qaYQAPsM1zcug0CZVf4328cdsWsl9dzHwIfk,7498
|
|
80
|
+
inference_models/models/perception_encoder/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
81
|
+
inference_models/models/perception_encoder/perception_encoder_pytorch.py,sha256=RBcMCQQnqsSV8YiEKqdQb1woTaRKAIQahcWIfhvMhxQ,7535
|
|
82
|
+
inference_models/models/perception_encoder/vision_encoder/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
83
|
+
inference_models/models/perception_encoder/vision_encoder/config.py,sha256=VLf_sLKCsGz1wIJAaEsRP6I6jMYHRgfrikH92QFfDSY,3500
|
|
84
|
+
inference_models/models/perception_encoder/vision_encoder/pe.py,sha256=6VL7tJMTt0YxKt20g8s2yq7BKmA8YsKUD7mhhqP9WEM,23709
|
|
85
|
+
inference_models/models/perception_encoder/vision_encoder/rope.py,sha256=gNOzAZzQ-3xqmJQgpnpPBNfWi-riqOe3BsSowJnvtLA,10502
|
|
86
|
+
inference_models/models/perception_encoder/vision_encoder/tokenizer.py,sha256=doV4hmh4bk1VnbvFlK4U5sY06Ky9Bx89Uti89RRoIj0,11542
|
|
87
|
+
inference_models/models/perception_encoder/vision_encoder/transforms.py,sha256=z5uc86siRWYSfXV5nL1B67p4hGUj35xSocNPq2LpTNI,922
|
|
88
|
+
inference_models/models/qwen25vl/__init__.py,sha256=ivA02JoS6FqnY-bi-cst9H7j7JmGtQPZQS3EIkF5DI8,58
|
|
89
|
+
inference_models/models/qwen25vl/qwen25vl_hf.py,sha256=mJp8_X82eU8e2dAJERuDuMHhbegN21HYDjQ_luRFcsI,10033
|
|
90
|
+
inference_models/models/resnet/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
91
|
+
inference_models/models/resnet/resnet_classification_onnx.py,sha256=Mv2aqQlRhKzMCJxrPDJ7u6hubONvJvhPfUQ3nt_xkqI,12679
|
|
92
|
+
inference_models/models/resnet/resnet_classification_torch.py,sha256=sxfhdyhidElNni-l-kSzNBqI4SMuMt6XrQieJsrUvEc,10795
|
|
93
|
+
inference_models/models/resnet/resnet_classification_trt.py,sha256=WvS2HtpDjNqTLmutcD5bXpYgWvOexFAcW5wRTDxeuuo,13528
|
|
94
|
+
inference_models/models/rfdetr/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
95
|
+
inference_models/models/rfdetr/backbone_builder.py,sha256=PqhljbZB2b88oJ0aSf4jJrEUcB2iHy1OEWMricY1e98,2866
|
|
96
|
+
inference_models/models/rfdetr/class_remapping.py,sha256=JXTBF7uCCmFHnjO4ffmagERk6JT-JO5HrHu-2Z0QKKM,1393
|
|
97
|
+
inference_models/models/rfdetr/common.py,sha256=oNk6T5rCrnvpW3QhBqzWo1zj7IoWSXeEvwI5WhaNGSA,4679
|
|
98
|
+
inference_models/models/rfdetr/default_labels.py,sha256=BAFZJwZLvvN8Ui4Nxc2ScbEhdiClo3AwFwV6Zd0vR8s,1849
|
|
99
|
+
inference_models/models/rfdetr/dinov2_with_windowed_attn.py,sha256=2wYXXVaqNK45dCowUGsHD63VXcRtZdPHGjv_V_yUd4w,54220
|
|
100
|
+
inference_models/models/rfdetr/misc.py,sha256=EhzPmWNA5yDJAMQN0E05i6e5BI6Scow53eVrpNc1Pus,666
|
|
101
|
+
inference_models/models/rfdetr/ms_deform_attn.py,sha256=IkrRLMel05bz0SLQGGK-vNJuwkBtiHeEp12Tc3Xqq8w,7304
|
|
102
|
+
inference_models/models/rfdetr/ms_deform_attn_func.py,sha256=j5GzCZjgG8bWPTBKpcNVG2SC9yUJ0HFPWAogKPck5HQ,2706
|
|
103
|
+
inference_models/models/rfdetr/position_encoding.py,sha256=aZAlbgh7QYjvd9c_5FTUv_FSie3UkHY22cnLIcs6od0,6104
|
|
104
|
+
inference_models/models/rfdetr/post_processor.py,sha256=Lq0jfmORWTBRLfPI9Hf3bQbigafivGuaLy9nosmKdxI,3071
|
|
105
|
+
inference_models/models/rfdetr/projector.py,sha256=JZHQxRTj_8lt0Bo2M2aq4bSKtOk9x3UyuLVgA8LYXWA,12684
|
|
106
|
+
inference_models/models/rfdetr/rfdetr_backbone_pytorch.py,sha256=e5K9rhKsGMAhU1UxfLib6ivgsxCHDWc7TrOVXeJ7B0A,14209
|
|
107
|
+
inference_models/models/rfdetr/rfdetr_base_pytorch.py,sha256=eFMmwLYhaZ8yNCoMhCkngTaU3cxiEdX_a-ueHt7bapA,28971
|
|
108
|
+
inference_models/models/rfdetr/rfdetr_instance_segmentation_onnx.py,sha256=3lFZLn6HdOtEHo1op_1UN9tvsA-f7T4xL3KPX3RarLg,8024
|
|
109
|
+
inference_models/models/rfdetr/rfdetr_instance_segmentation_pytorch.py,sha256=YUfTRAo_2GOVOpbu6Csy8NWKraDvZXK6L-bfuYxB22s,14926
|
|
110
|
+
inference_models/models/rfdetr/rfdetr_instance_segmentation_trt.py,sha256=qieNBjkhmbbKQyvUGTdtuP93Hn3rDFqYx9PoWPMkE6k,8398
|
|
111
|
+
inference_models/models/rfdetr/rfdetr_object_detection_onnx.py,sha256=MA7-ww9h9YRtx7WlPfgElZhQmnPpZL8TOiNfUogmUy8,9765
|
|
112
|
+
inference_models/models/rfdetr/rfdetr_object_detection_pytorch.py,sha256=UMYQvXayNZnyWVNIN3V9mcXbooqj5xsxPHb3S94pN6A,19329
|
|
113
|
+
inference_models/models/rfdetr/rfdetr_object_detection_trt.py,sha256=XZGRooT8qywjTVt-6IUX5oogHezu9QfD_rcGeyzCeHQ,10423
|
|
114
|
+
inference_models/models/rfdetr/segmentation_head.py,sha256=qg0m09xJVQ_DP5lYx_IkYbuC3UsLcv915cMQx65Alio,10214
|
|
115
|
+
inference_models/models/rfdetr/transformer.py,sha256=aYbE9x6QBA5T56pAE9lQ_lgMreG3INdAT5LaywOto9c,27502
|
|
116
|
+
inference_models/models/roboflow_instant/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
117
|
+
inference_models/models/roboflow_instant/roboflow_instant_hf.py,sha256=UN5_wIhCZZjBmPKzcP0MDAQbEizcYSnufAga2Uru3cM,5309
|
|
118
|
+
inference_models/models/sam/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
119
|
+
inference_models/models/sam/cache.py,sha256=ZtQnhLBOeyVnaknMLErueovcFheaM1cbXtg5pt2h_i8,4742
|
|
120
|
+
inference_models/models/sam/entities.py,sha256=72WR3mVkxGJeRoZbcObuS5Ay6Mw8cqMtqqfyKm6DLfQ,584
|
|
121
|
+
inference_models/models/sam/sam_torch.py,sha256=gOPJT4Fi-kqrhFZrlblm1MZ17rg2MD9cIWqJFplK9Xo,29397
|
|
122
|
+
inference_models/models/sam2/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
123
|
+
inference_models/models/sam2/cache.py,sha256=2tNkn_WGaZnhmc24cMZd3zmGP2nYFzW_IVEh0A-61Gc,5609
|
|
124
|
+
inference_models/models/sam2/entities.py,sha256=33TZpSv1bv5MWkZu77ksgp0RGO-fz93_J5NqtiweAMQ,1154
|
|
125
|
+
inference_models/models/sam2/sam2_torch.py,sha256=8yaXHnYRklOrpiP6f8WuJzFnaoNWlt0n6R0kM3Kb4JM,38787
|
|
126
|
+
inference_models/models/sam2_rt/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
127
|
+
inference_models/models/sam2_rt/sam2_pytorch.py,sha256=-hV_osk_RCzRj1delAYaTFneZ-nUP3lGNwtTmX5Uzm8,4482
|
|
128
|
+
inference_models/models/smolvlm/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
129
|
+
inference_models/models/smolvlm/smolvlm_hf.py,sha256=nwAFziRoxlK-pwnF6vIwB331IIW_H57v8t-Ai5h6fH0,8903
|
|
130
|
+
inference_models/models/trocr/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
131
|
+
inference_models/models/trocr/trocr_hf.py,sha256=rFdTySKEH5pyBneQrGZA4_4Ix9-SZ1zp8YZ7ZFAmeUQ,1786
|
|
132
|
+
inference_models/models/vit/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
133
|
+
inference_models/models/vit/vit_classification_huggingface.py,sha256=Lq5Z9X4i4ghpxF0QzLQduJ5n2r6OulOCLyJg0x4kLG4,11027
|
|
134
|
+
inference_models/models/vit/vit_classification_onnx.py,sha256=ml05Mq4YpC5Ibca5AX_y54c43Q87UU5lKev82BpHYT4,12473
|
|
135
|
+
inference_models/models/vit/vit_classification_trt.py,sha256=1R8jJb6tMFejCEJje2pzoFGqoj4FYmQwJSu1OiTvLBY,13325
|
|
136
|
+
inference_models/models/yolact/__init__.py,sha256=wMN3BTunf5CFJH_sV_LIgMUufYfiFwVZUjbaBbdlq1M,33
|
|
137
|
+
inference_models/models/yolact/yolact_instance_segmentation_onnx.py,sha256=SK3-uC4aJaZtfE89KCD1E9hmySIx50hIT8py0pu_YnM,13272
|
|
138
|
+
inference_models/models/yolact/yolact_instance_segmentation_trt.py,sha256=APh0QS4WELptxpaHBiqF6PYZhouks7gTZcoLxJEeE2s,13770
|
|
139
|
+
inference_models/models/yolo_world/__init__.py,sha256=sQfPkM6WqydxTa9ORxqCaSjsVUjCWSIh8_Fp6VcWudM,73
|
|
140
|
+
inference_models/models/yolonas/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
141
|
+
inference_models/models/yolonas/nms.py,sha256=kYLONdjr8vp9yjjpst8jiD_16qMM2l3ArXd35bL_cQo,1443
|
|
142
|
+
inference_models/models/yolonas/yolonas_object_detection_onnx.py,sha256=-OqZQs8pvke65teh09H6AsjQ4z8kxLa1kKK98x45ZJQ,8047
|
|
143
|
+
inference_models/models/yolonas/yolonas_object_detection_trt.py,sha256=rHz5Z9QSkUnyDqKohDZnArknytybDBjKsCYudFBPH2o,8632
|
|
144
|
+
inference_models/models/yolov10/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
145
|
+
inference_models/models/yolov10/yolov10_object_detection_onnx.py,sha256=dj4er3d2p7vP5AYaTNEMi0Dy9qd2AQvukN1lf8BlOLk,7272
|
|
146
|
+
inference_models/models/yolov10/yolov10_object_detection_trt.py,sha256=J71uE4xCEUtoFs8fXwDkIfVFsdHTckR5gA8zTE76Pxs,7893
|
|
147
|
+
inference_models/models/yolov11/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
148
|
+
inference_models/models/yolov11/yolov11_onnx.py,sha256=9wYERJDCjxa6BxotjmOCKN9Nf2tEVnCcGjCmJ_cl-bI,783
|
|
149
|
+
inference_models/models/yolov11/yolov11_torch_script.py,sha256=i1CFesljr4t3OtWVGk_WzPEJvbvDxnQgL0EFj4vH8Ww,699
|
|
150
|
+
inference_models/models/yolov11/yolov11_trt.py,sha256=vhg7MzfEQ7rpGUVvQI_MmS25GwR8UjwVKgRHmLTOu2M,588
|
|
151
|
+
inference_models/models/yolov12/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
152
|
+
inference_models/models/yolov12/yolov12_onnx.py,sha256=sSW8U7z_hHq2Rq_vdG2qbrPikNEaAm29M6rs4X_CSeo,188
|
|
153
|
+
inference_models/models/yolov12/yolov12_torch_script.py,sha256=pH7dFi1JgFCt8oIBe0gmeBvsdMWDU2R1z4icWVBkToI,217
|
|
154
|
+
inference_models/models/yolov12/yolov12_trt.py,sha256=vLhn2ZJfAmgZW7UiowmZosUMo-XhIWrZeXhWbT0Rx-c,184
|
|
155
|
+
inference_models/models/yolov5/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
156
|
+
inference_models/models/yolov5/nms.py,sha256=3uu8JpBBZTUkaSfcVziYvQ3QhXTrfDmOCs2adkR8D6U,3458
|
|
157
|
+
inference_models/models/yolov5/yolov5_instance_segmentation_onnx.py,sha256=eVKQFVUDz8B1bWdkBYG02yzX0ya6-WDKQS6l5Cr5Ico,8873
|
|
158
|
+
inference_models/models/yolov5/yolov5_instance_segmentation_trt.py,sha256=E285z4QtO1I6Zud18w-X-9ts_XMIUU8dyxEELqrqlGA,9724
|
|
159
|
+
inference_models/models/yolov5/yolov5_object_detection_onnx.py,sha256=TTwAZYVWBz8VXug0jHKxWOKB8vMQrvhnb0opXDYg67M,7452
|
|
160
|
+
inference_models/models/yolov5/yolov5_object_detection_trt.py,sha256=bgp4ubTxwoi14QXhXuNsILdHKFsZ6fikjg9gdP76ViA,8007
|
|
161
|
+
inference_models/models/yolov7/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
162
|
+
inference_models/models/yolov7/yolov7_instance_segmentation_onnx.py,sha256=GhZx8Dx_QEVFU9A1prydFP2TRPXobGIE72dt6v_B1XA,8874
|
|
163
|
+
inference_models/models/yolov7/yolov7_instance_segmentation_trt.py,sha256=YXOG78aZ_3koNKu_kyUcqifXaKyud15a-P6-PRI4Lvw,9478
|
|
164
|
+
inference_models/models/yolov8/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
165
|
+
inference_models/models/yolov8/yolov8_classification_onnx.py,sha256=fMi-mkdzsUi9eWfcdn5_iEQ-D5LqxKAJaHzC8tcj0wo,7111
|
|
166
|
+
inference_models/models/yolov8/yolov8_instance_segmentation_onnx.py,sha256=CL2tNH09TjO-PCIONBbALt8fy113HLNz2f0oY_x3dS4,9425
|
|
167
|
+
inference_models/models/yolov8/yolov8_instance_segmentation_torch_script.py,sha256=q6qZdVX7TimnyqzBDlk0uFykWZKji16S-gyvlCOmhrs,7701
|
|
168
|
+
inference_models/models/yolov8/yolov8_instance_segmentation_trt.py,sha256=i7tOfSdKVGANCKXJx0_0E03yVzrtK2CrGuOpKbyx_ks,10178
|
|
169
|
+
inference_models/models/yolov8/yolov8_key_points_detection_onnx.py,sha256=IKQANr9fxIEwnQ7K7QNi4NYKtiC0yGiU_aduBUGkXLs,10612
|
|
170
|
+
inference_models/models/yolov8/yolov8_key_points_detection_torch_script.py,sha256=986Bx2uSG9eBEyaBv8G50OOZoT3trrls7pzguttFTgo,8552
|
|
171
|
+
inference_models/models/yolov8/yolov8_key_points_detection_trt.py,sha256=xl0ha3onxR8BqtnLTuUEDdaJIsJpyxBL8FWMYT0TXis,11139
|
|
172
|
+
inference_models/models/yolov8/yolov8_object_detection_onnx.py,sha256=oJR1OmiWUemO5RwaoePO-IwXk1aQsxWV0Bhhd0MvoeM,8251
|
|
173
|
+
inference_models/models/yolov8/yolov8_object_detection_torch_script.py,sha256=M3R5fcrTRa0xkXCna9OkiOqthReANi7Qm2TuI-5Gs4o,6104
|
|
174
|
+
inference_models/models/yolov8/yolov8_object_detection_trt.py,sha256=pt0dHffNR6KVFXgy05EKwcMXP1FJfMqoChiYkNJ-gTk,8478
|
|
175
|
+
inference_models/models/yolov9/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
176
|
+
inference_models/models/yolov9/yolov9_onnx.py,sha256=zV7LK3CtSVR1R9zKJjCdVUvmCbDNg66PDI1mgorlk74,187
|
|
177
|
+
inference_models/models/yolov9/yolov9_torch_script.py,sha256=YmOE3ioUrLvU_SyPQxfIrErNrbBUxDUcyHHPx6IlW7c,216
|
|
178
|
+
inference_models/models/yolov9/yolov9_trt.py,sha256=l1PUidw005b0hUF2VAd45GV6ICSwxK48Txt_P_InVIg,183
|
|
179
|
+
inference_models/runtime_introspection/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
180
|
+
inference_models/runtime_introspection/core.py,sha256=9_XaCekjX1OROlSNsZ4q68iImhn6pbxqtJE0a_4GDk4,13230
|
|
181
|
+
inference_models/utils/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
182
|
+
inference_models/utils/download.py,sha256=YoVvZBIJAiJu2OqBJfDiYBcih8M4tAGUD9QFB5w7Dl4,21950
|
|
183
|
+
inference_models/utils/environment.py,sha256=koaOp-b6ovv_mCLVhO7wzIX26Pg9wi0oBj6vWoOhP4I,853
|
|
184
|
+
inference_models/utils/file_system.py,sha256=Z5s2AmNeTVvzM6OHy8fJPsboqQFttbzTMpUV7Ibg6SU,1405
|
|
185
|
+
inference_models/utils/hashing.py,sha256=Gb0ISj6cc32aZ94oI9BwtIEw6AYOSJaL2YiGycFQVRg,194
|
|
186
|
+
inference_models/utils/imports.py,sha256=ZqsVmdVVdfl3V1Ke7HRnXv0CIUE-6J0RlGCef9EfdcA,1514
|
|
187
|
+
inference_models/utils/onnx_introspection.py,sha256=sxQCdDWXtPokUgMhK4mN_Hxiv4ea85aGAnPnt63Z6q4,460
|
|
188
|
+
inference_models/weights_providers/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
189
|
+
inference_models/weights_providers/core.py,sha256=ZCm7R-svrgaiOZ3AXpGA5OsYcDwhtwGRN2la5mpG_GA,715
|
|
190
|
+
inference_models/weights_providers/entities.py,sha256=hI-tQ7BkdabnBfMko2wYo76jRGWXJUWlSTwq2OQ4RDM,5834
|
|
191
|
+
inference_models/weights_providers/roboflow.py,sha256=6tXzqolMnQLIz5vR_eXQsZ1DmZ9UMPlimddZLQFA_8Y,25358
|
|
192
|
+
inference_models-0.18.3.dist-info/METADATA,sha256=QCLAbiUACa97jaTjKGGDVTq-Maa31PS9GtIqjlxldgU,20940
|
|
193
|
+
inference_models-0.18.3.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
|
|
194
|
+
inference_models-0.18.3.dist-info/top_level.txt,sha256=128IRjtd3k9I4c7i-eErAE2oOlOIF6-6iRW7tHm7VxE,17
|
|
195
|
+
inference_models-0.18.3.dist-info/RECORD,,
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
inference_models
|