@camstack/addon-vision 0.1.6 → 0.1.9

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (136) hide show
  1. package/dist/addons/animal-classifier/index.d.mts +30 -0
  2. package/dist/addons/animal-classifier/index.d.ts +30 -0
  3. package/dist/addons/animal-classifier/index.js +822 -999
  4. package/dist/addons/animal-classifier/index.js.map +1 -1
  5. package/dist/addons/animal-classifier/index.mjs +7 -242
  6. package/dist/addons/animal-classifier/index.mjs.map +1 -1
  7. package/dist/addons/audio-classification/index.d.mts +36 -0
  8. package/dist/addons/audio-classification/index.d.ts +36 -0
  9. package/dist/addons/audio-classification/index.js +378 -501
  10. package/dist/addons/audio-classification/index.js.map +1 -1
  11. package/dist/addons/audio-classification/index.mjs +4 -224
  12. package/dist/addons/audio-classification/index.mjs.map +1 -1
  13. package/dist/addons/bird-global-classifier/index.d.mts +31 -0
  14. package/dist/addons/bird-global-classifier/index.d.ts +31 -0
  15. package/dist/addons/bird-global-classifier/index.js +825 -1002
  16. package/dist/addons/bird-global-classifier/index.js.map +1 -1
  17. package/dist/addons/bird-global-classifier/index.mjs +7 -248
  18. package/dist/addons/bird-global-classifier/index.mjs.map +1 -1
  19. package/dist/addons/bird-nabirds-classifier/index.d.mts +33 -0
  20. package/dist/addons/bird-nabirds-classifier/index.d.ts +33 -0
  21. package/dist/addons/bird-nabirds-classifier/index.js +825 -1002
  22. package/dist/addons/bird-nabirds-classifier/index.js.map +1 -1
  23. package/dist/addons/bird-nabirds-classifier/index.mjs +7 -289
  24. package/dist/addons/bird-nabirds-classifier/index.mjs.map +1 -1
  25. package/dist/addons/face-detection/index.d.mts +29 -0
  26. package/dist/addons/face-detection/index.d.ts +29 -0
  27. package/dist/addons/face-detection/index.js +934 -1196
  28. package/dist/addons/face-detection/index.js.map +1 -1
  29. package/dist/addons/face-detection/index.mjs +7 -227
  30. package/dist/addons/face-detection/index.mjs.map +1 -1
  31. package/dist/addons/face-recognition/index.d.mts +29 -0
  32. package/dist/addons/face-recognition/index.d.ts +29 -0
  33. package/dist/addons/face-recognition/index.js +807 -1003
  34. package/dist/addons/face-recognition/index.js.map +1 -1
  35. package/dist/addons/face-recognition/index.mjs +6 -197
  36. package/dist/addons/face-recognition/index.mjs.map +1 -1
  37. package/dist/addons/motion-detection/index.d.mts +28 -0
  38. package/dist/addons/motion-detection/index.d.ts +28 -0
  39. package/dist/addons/motion-detection/index.js +111 -214
  40. package/dist/addons/motion-detection/index.js.map +1 -1
  41. package/dist/addons/motion-detection/index.mjs +9 -12
  42. package/dist/addons/motion-detection/index.mjs.map +1 -1
  43. package/dist/addons/object-detection/index.d.mts +31 -0
  44. package/dist/addons/object-detection/index.d.ts +31 -0
  45. package/dist/addons/object-detection/index.js +1082 -1287
  46. package/dist/addons/object-detection/index.js.map +1 -1
  47. package/dist/addons/object-detection/index.mjs +7 -373
  48. package/dist/addons/object-detection/index.mjs.map +1 -1
  49. package/dist/addons/plate-detection/index.d.mts +30 -0
  50. package/dist/addons/plate-detection/index.d.ts +30 -0
  51. package/dist/addons/plate-detection/index.js +868 -1075
  52. package/dist/addons/plate-detection/index.js.map +1 -1
  53. package/dist/addons/plate-detection/index.mjs +7 -230
  54. package/dist/addons/plate-detection/index.mjs.map +1 -1
  55. package/dist/addons/plate-recognition/index.d.mts +31 -0
  56. package/dist/addons/plate-recognition/index.d.ts +31 -0
  57. package/dist/addons/plate-recognition/index.js +505 -684
  58. package/dist/addons/plate-recognition/index.js.map +1 -1
  59. package/dist/addons/plate-recognition/index.mjs +5 -244
  60. package/dist/addons/plate-recognition/index.mjs.map +1 -1
  61. package/dist/addons/segmentation-refiner/index.d.mts +30 -0
  62. package/dist/addons/segmentation-refiner/index.d.ts +30 -0
  63. package/dist/addons/segmentation-refiner/index.js +790 -967
  64. package/dist/addons/segmentation-refiner/index.js.map +1 -1
  65. package/dist/addons/segmentation-refiner/index.mjs +17 -21
  66. package/dist/addons/segmentation-refiner/index.mjs.map +1 -1
  67. package/dist/addons/vehicle-classifier/index.d.mts +31 -0
  68. package/dist/addons/vehicle-classifier/index.d.ts +31 -0
  69. package/dist/addons/vehicle-classifier/index.js +410 -581
  70. package/dist/addons/vehicle-classifier/index.js.map +1 -1
  71. package/dist/addons/vehicle-classifier/index.mjs +16 -20
  72. package/dist/addons/vehicle-classifier/index.mjs.map +1 -1
  73. package/dist/chunk-22BHCDT5.mjs +101 -0
  74. package/dist/{chunk-WG66JYYW.mjs.map → chunk-22BHCDT5.mjs.map} +1 -1
  75. package/dist/chunk-2IOKI4ES.mjs +335 -0
  76. package/dist/{chunk-PIFS7AIT.mjs.map → chunk-2IOKI4ES.mjs.map} +1 -1
  77. package/dist/chunk-7DYHXUPZ.mjs +36 -0
  78. package/dist/{chunk-BS4DKYGN.mjs.map → chunk-7DYHXUPZ.mjs.map} +1 -1
  79. package/dist/chunk-BJTO5JO5.mjs +11 -0
  80. package/dist/chunk-BP7H4NFS.mjs +412 -0
  81. package/dist/{chunk-MGT6RUVX.mjs.map → chunk-BP7H4NFS.mjs.map} +1 -1
  82. package/dist/chunk-BR2FPGOX.mjs +98 -0
  83. package/dist/{chunk-YYDM6V2F.mjs.map → chunk-BR2FPGOX.mjs.map} +1 -1
  84. package/dist/chunk-D6WEHN33.mjs +276 -0
  85. package/dist/chunk-D6WEHN33.mjs.map +1 -0
  86. package/dist/chunk-DRYFGARD.mjs +289 -0
  87. package/dist/chunk-DRYFGARD.mjs.map +1 -0
  88. package/dist/chunk-DUN6XU3N.mjs +72 -0
  89. package/dist/{chunk-XD7WGXHZ.mjs.map → chunk-DUN6XU3N.mjs.map} +1 -1
  90. package/dist/chunk-ESLHNWWE.mjs +387 -0
  91. package/dist/chunk-ESLHNWWE.mjs.map +1 -0
  92. package/dist/chunk-JUQEW6ON.mjs +256 -0
  93. package/dist/chunk-JUQEW6ON.mjs.map +1 -0
  94. package/dist/chunk-KUO2BVFY.mjs +90 -0
  95. package/dist/{chunk-DE7I3VHO.mjs.map → chunk-KUO2BVFY.mjs.map} +1 -1
  96. package/dist/chunk-R5J3WAUI.mjs +645 -0
  97. package/dist/chunk-R5J3WAUI.mjs.map +1 -0
  98. package/dist/chunk-XZ6ZMXXU.mjs +39 -0
  99. package/dist/{chunk-K36R6HWY.mjs.map → chunk-XZ6ZMXXU.mjs.map} +1 -1
  100. package/dist/chunk-YPU4WTXZ.mjs +269 -0
  101. package/dist/chunk-YPU4WTXZ.mjs.map +1 -0
  102. package/dist/chunk-YUCD2TFH.mjs +242 -0
  103. package/dist/chunk-YUCD2TFH.mjs.map +1 -0
  104. package/dist/chunk-ZTJENCFC.mjs +379 -0
  105. package/dist/chunk-ZTJENCFC.mjs.map +1 -0
  106. package/dist/chunk-ZWYXXCXP.mjs +248 -0
  107. package/dist/chunk-ZWYXXCXP.mjs.map +1 -0
  108. package/dist/index.d.mts +183 -0
  109. package/dist/index.d.ts +183 -0
  110. package/dist/index.js +3930 -4449
  111. package/dist/index.js.map +1 -1
  112. package/dist/index.mjs +250 -2698
  113. package/dist/index.mjs.map +1 -1
  114. package/package.json +5 -5
  115. package/dist/chunk-2YMA6QOV.mjs +0 -193
  116. package/dist/chunk-2YMA6QOV.mjs.map +0 -1
  117. package/dist/chunk-3IIFBJCD.mjs +0 -45
  118. package/dist/chunk-BS4DKYGN.mjs +0 -48
  119. package/dist/chunk-DE7I3VHO.mjs +0 -106
  120. package/dist/chunk-F6D2OZ36.mjs +0 -89
  121. package/dist/chunk-F6D2OZ36.mjs.map +0 -1
  122. package/dist/chunk-GAOIFQDX.mjs +0 -59
  123. package/dist/chunk-GAOIFQDX.mjs.map +0 -1
  124. package/dist/chunk-HUIX2XVR.mjs +0 -159
  125. package/dist/chunk-HUIX2XVR.mjs.map +0 -1
  126. package/dist/chunk-K36R6HWY.mjs +0 -51
  127. package/dist/chunk-MBTAI3WE.mjs +0 -78
  128. package/dist/chunk-MBTAI3WE.mjs.map +0 -1
  129. package/dist/chunk-MGT6RUVX.mjs +0 -423
  130. package/dist/chunk-PIFS7AIT.mjs +0 -446
  131. package/dist/chunk-WG66JYYW.mjs +0 -116
  132. package/dist/chunk-XD7WGXHZ.mjs +0 -82
  133. package/dist/chunk-YYDM6V2F.mjs +0 -113
  134. package/dist/chunk-ZK7P3TZN.mjs +0 -286
  135. package/dist/chunk-ZK7P3TZN.mjs.map +0 -1
  136. /package/dist/{chunk-3IIFBJCD.mjs.map → chunk-BJTO5JO5.mjs.map} +0 -0
@@ -1 +0,0 @@
1
- {"version":3,"sources":["../src/shared/postprocess/yamnet.ts","../src/catalogs/audio-classification-models.ts"],"sourcesContent":["export interface AudioClassification {\n readonly className: string\n readonly score: number\n}\n\n/** Average YAMNET scores across frames, return top classes above threshold */\nexport function yamnetPostprocess(\n output: Float32Array,\n numFrames: number,\n numClasses: number,\n classNames: readonly string[],\n minScore: number,\n): AudioClassification[] {\n // Average across frames\n const avgScores = new Float32Array(numClasses)\n for (let f = 0; f < numFrames; f++) {\n for (let c = 0; c < numClasses; c++) {\n const prev = avgScores[c] ?? 0\n avgScores[c] = prev + (output[f * numClasses + c] ?? 0)\n }\n }\n if (numFrames > 0) {\n for (let c = 0; c < numClasses; c++) {\n const val = avgScores[c] ?? 0\n avgScores[c] = val / numFrames\n }\n }\n\n // Collect classes above threshold\n const results: AudioClassification[] = []\n for (let c = 0; c < numClasses; c++) {\n const score = avgScores[c]!\n if (score >= minScore) {\n results.push({\n className: classNames[c] ?? String(c),\n score,\n })\n }\n }\n\n // Sort descending by score\n return results.sort((a, b) => b.score - a.score)\n}\n","import type { ModelCatalogEntry, LabelDefinition } from '@camstack/types'\nimport { hfModelUrl } from '@camstack/types'\n\nconst HF_REPO = 'camstack/camstack-models'\n\nconst AUDIO_LABELS: readonly LabelDefinition[] = [\n { id: 'audio', name: 'Audio Event' },\n] as const\n\nexport const AUDIO_CLASSIFICATION_MODELS: readonly ModelCatalogEntry[] = [\n {\n id: 'yamnet',\n name: 'YAMNet',\n description: 'YAMNet — audio event classification from raw waveform',\n inputSize: { width: 1, height: 16000 },\n labels: AUDIO_LABELS,\n formats: {\n onnx: {\n url: hfModelUrl(HF_REPO, 'audioClassification/yamnet/onnx/camstack-yamnet.onnx'),\n sizeMB: 15,\n },\n openvino: {\n url: hfModelUrl(HF_REPO, 'audioClassification/yamnet/openvino/camstack-yamnet.xml'),\n sizeMB: 8,\n },\n },\n },\n] as const\n"],"mappings":";;;;;;;;;;AAMA,YAAA,oBAAA;AAAA,aAAgB,kBACd,QACA,WACA,YACA,YACA,UAAgB;AAGhB,YAAM,YAAY,IAAI,aAAa,UAAU;AAC7C,eAAS,IAAI,GAAG,IAAI,WAAW,KAAK;AAClC,iBAAS,IAAI,GAAG,IAAI,YAAY,KAAK;AACnC,gBAAM,OAAO,UAAU,CAAC,KAAK;AAC7B,oBAAU,CAAC,IAAI,QAAQ,OAAO,IAAI,aAAa,CAAC,KAAK;QACvD;MACF;AACA,UAAI,YAAY,GAAG;AACjB,iBAAS,IAAI,GAAG,IAAI,YAAY,KAAK;AACnC,gBAAM,MAAM,UAAU,CAAC,KAAK;AAC5B,oBAAU,CAAC,IAAI,MAAM;QACvB;MACF;AAGA,YAAM,UAAiC,CAAA;AACvC,eAAS,IAAI,GAAG,IAAI,YAAY,KAAK;AACnC,cAAM,QAAQ,UAAU,CAAC;AACzB,YAAI,SAAS,UAAU;AACrB,kBAAQ,KAAK;YACX,WAAW,WAAW,CAAC,KAAK,OAAO,CAAC;YACpC;WACD;QACH;MACF;AAGA,aAAO,QAAQ,KAAK,CAAC,GAAG,MAAM,EAAE,QAAQ,EAAE,KAAK;IACjD;;;;;;;;;;ACzCA,QAAA,UAAA,UAAA,iBAAA;AAEA,QAAM,UAAU;AAEhB,QAAM,eAA2C;MAC/C,EAAE,IAAI,SAAS,MAAM,cAAa;;AAGvB,YAAA,8BAA4D;MACvE;QACE,IAAI;QACJ,MAAM;QACN,aAAa;QACb,WAAW,EAAE,OAAO,GAAG,QAAQ,KAAK;QACpC,QAAQ;QACR,SAAS;UACP,MAAM;YACJ,MAAK,GAAA,QAAA,YAAW,SAAS,sDAAsD;YAC/E,QAAQ;;UAEV,UAAU;YACR,MAAK,GAAA,QAAA,YAAW,SAAS,yDAAyD;YAClF,QAAQ;;;;;;;","names":[]}
@@ -1,423 +0,0 @@
1
- import {
2
- __commonJS,
3
- __require
4
- } from "./chunk-3IIFBJCD.mjs";
5
-
6
- // src/catalogs/object-detection-models.js
7
- var require_object_detection_models = __commonJS({
8
- "src/catalogs/object-detection-models.js"(exports) {
9
- "use strict";
10
- Object.defineProperty(exports, "__esModule", { value: true });
11
- exports.OBJECT_DETECTION_MODELS = exports.MLPACKAGE_FILES = void 0;
12
- var types_1 = __require("@camstack/types");
13
- var HF_REPO = "camstack/camstack-models";
14
- exports.MLPACKAGE_FILES = [
15
- "Manifest.json",
16
- "Data/com.apple.CoreML/model.mlmodel",
17
- "Data/com.apple.CoreML/weights/weight.bin"
18
- ];
19
- exports.OBJECT_DETECTION_MODELS = [
20
- // ── YOLOv8 ──────────────────────────────────────────────────────
21
- {
22
- id: "yolov8n",
23
- name: "YOLOv8 Nano",
24
- description: "YOLOv8 Nano \u2014 fastest, smallest object detection model",
25
- inputSize: { width: 640, height: 640 },
26
- labels: types_1.COCO_80_LABELS,
27
- formats: {
28
- onnx: {
29
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/onnx/camstack-yolov8n.onnx"),
30
- sizeMB: 12
31
- },
32
- coreml: {
33
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/coreml/camstack-yolov8n.mlpackage"),
34
- sizeMB: 6,
35
- isDirectory: true,
36
- files: exports.MLPACKAGE_FILES,
37
- runtimes: ["python"]
38
- },
39
- openvino: {
40
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/openvino/camstack-yolov8n.xml"),
41
- sizeMB: 7,
42
- runtimes: ["python"]
43
- },
44
- tflite: {
45
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/tflite/camstack-yolov8n_float32.tflite"),
46
- sizeMB: 12,
47
- runtimes: ["python"]
48
- }
49
- }
50
- },
51
- {
52
- id: "yolov8s",
53
- name: "YOLOv8 Small",
54
- description: "YOLOv8 Small \u2014 balanced speed and accuracy",
55
- inputSize: { width: 640, height: 640 },
56
- labels: types_1.COCO_80_LABELS,
57
- formats: {
58
- onnx: {
59
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/onnx/camstack-yolov8s.onnx"),
60
- sizeMB: 43
61
- },
62
- coreml: {
63
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/coreml/camstack-yolov8s.mlpackage"),
64
- sizeMB: 21,
65
- isDirectory: true,
66
- files: exports.MLPACKAGE_FILES,
67
- runtimes: ["python"]
68
- },
69
- openvino: {
70
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/openvino/camstack-yolov8s.xml"),
71
- sizeMB: 22,
72
- runtimes: ["python"]
73
- },
74
- tflite: {
75
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/tflite/camstack-yolov8s_float32.tflite"),
76
- sizeMB: 43,
77
- runtimes: ["python"]
78
- }
79
- }
80
- },
81
- {
82
- id: "yolov8s-relu",
83
- name: "YOLOv8 Small ReLU",
84
- description: "YOLOv8 Small with ReLU activation \u2014 better hardware compatibility",
85
- inputSize: { width: 640, height: 640 },
86
- labels: types_1.COCO_80_LABELS,
87
- formats: {
88
- onnx: {
89
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/onnx/camstack-yolov8s-relu.onnx"),
90
- sizeMB: 43
91
- }
92
- }
93
- },
94
- {
95
- id: "yolov8m",
96
- name: "YOLOv8 Medium",
97
- description: "YOLOv8 Medium \u2014 higher accuracy, moderate size",
98
- inputSize: { width: 640, height: 640 },
99
- labels: types_1.COCO_80_LABELS,
100
- formats: {
101
- onnx: {
102
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/onnx/camstack-yolov8m.onnx"),
103
- sizeMB: 99
104
- },
105
- coreml: {
106
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/coreml/camstack-yolov8m.mlpackage"),
107
- sizeMB: 49,
108
- isDirectory: true,
109
- files: exports.MLPACKAGE_FILES,
110
- runtimes: ["python"]
111
- },
112
- openvino: {
113
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/openvino/camstack-yolov8m.xml"),
114
- sizeMB: 50,
115
- runtimes: ["python"]
116
- },
117
- tflite: {
118
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/tflite/camstack-yolov8m_float32.tflite"),
119
- sizeMB: 99,
120
- runtimes: ["python"]
121
- }
122
- }
123
- },
124
- {
125
- id: "yolov8l",
126
- name: "YOLOv8 Large",
127
- description: "YOLOv8 Large \u2014 high-accuracy large model",
128
- inputSize: { width: 640, height: 640 },
129
- labels: types_1.COCO_80_LABELS,
130
- formats: {
131
- onnx: {
132
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/onnx/camstack-yolov8l.onnx"),
133
- sizeMB: 167
134
- },
135
- coreml: {
136
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/coreml/camstack-yolov8l.mlpackage"),
137
- sizeMB: 83,
138
- isDirectory: true,
139
- files: exports.MLPACKAGE_FILES,
140
- runtimes: ["python"]
141
- },
142
- openvino: {
143
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/openvino/camstack-yolov8l.xml"),
144
- sizeMB: 84,
145
- runtimes: ["python"]
146
- }
147
- }
148
- },
149
- {
150
- id: "yolov8x",
151
- name: "YOLOv8 Extra-Large",
152
- description: "YOLOv8 Extra-Large \u2014 maximum accuracy",
153
- inputSize: { width: 640, height: 640 },
154
- labels: types_1.COCO_80_LABELS,
155
- formats: {
156
- onnx: {
157
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/onnx/camstack-yolov8x.onnx"),
158
- sizeMB: 260
159
- },
160
- coreml: {
161
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/coreml/camstack-yolov8x.mlpackage"),
162
- sizeMB: 130,
163
- isDirectory: true,
164
- files: exports.MLPACKAGE_FILES,
165
- runtimes: ["python"]
166
- },
167
- openvino: {
168
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov8/openvino/camstack-yolov8x.xml"),
169
- sizeMB: 131,
170
- runtimes: ["python"]
171
- }
172
- }
173
- },
174
- // ── YOLOv9 ──────────────────────────────────────────────────────
175
- {
176
- id: "yolov9t",
177
- name: "YOLOv9 Tiny",
178
- description: "YOLOv9 Tiny \u2014 ultra-lightweight next-gen detector",
179
- inputSize: { width: 640, height: 640 },
180
- labels: types_1.COCO_80_LABELS,
181
- formats: {
182
- onnx: {
183
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov9/onnx/camstack-yolov9t.onnx"),
184
- sizeMB: 8
185
- },
186
- coreml: {
187
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov9/coreml/camstack-yolov9t.mlpackage"),
188
- sizeMB: 4,
189
- isDirectory: true,
190
- files: exports.MLPACKAGE_FILES,
191
- runtimes: ["python"]
192
- },
193
- openvino: {
194
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov9/openvino/camstack-yolov9t.xml"),
195
- sizeMB: 6,
196
- runtimes: ["python"]
197
- },
198
- tflite: {
199
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov9/tflite/camstack-yolov9t_float32.tflite"),
200
- sizeMB: 8,
201
- runtimes: ["python"]
202
- }
203
- }
204
- },
205
- {
206
- id: "yolov9s",
207
- name: "YOLOv9 Small",
208
- description: "YOLOv9 Small \u2014 improved efficiency over YOLOv8s",
209
- inputSize: { width: 640, height: 640 },
210
- labels: types_1.COCO_80_LABELS,
211
- formats: {
212
- onnx: {
213
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov9/onnx/camstack-yolov9s.onnx"),
214
- sizeMB: 28
215
- },
216
- coreml: {
217
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov9/coreml/camstack-yolov9s.mlpackage"),
218
- sizeMB: 14,
219
- isDirectory: true,
220
- files: exports.MLPACKAGE_FILES,
221
- runtimes: ["python"]
222
- },
223
- openvino: {
224
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov9/openvino/camstack-yolov9s.xml"),
225
- sizeMB: 16,
226
- runtimes: ["python"]
227
- },
228
- tflite: {
229
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov9/tflite/camstack-yolov9s_float32.tflite"),
230
- sizeMB: 28,
231
- runtimes: ["python"]
232
- }
233
- }
234
- },
235
- {
236
- id: "yolov9c",
237
- name: "YOLOv9 C",
238
- description: "YOLOv9 C \u2014 high-accuracy compact model",
239
- inputSize: { width: 640, height: 640 },
240
- labels: types_1.COCO_80_LABELS,
241
- formats: {
242
- onnx: {
243
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov9/onnx/camstack-yolov9c.onnx"),
244
- sizeMB: 97
245
- },
246
- coreml: {
247
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov9/coreml/camstack-yolov9c.mlpackage"),
248
- sizeMB: 48,
249
- isDirectory: true,
250
- files: exports.MLPACKAGE_FILES,
251
- runtimes: ["python"]
252
- },
253
- openvino: {
254
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov9/openvino/camstack-yolov9c.xml"),
255
- sizeMB: 49,
256
- runtimes: ["python"]
257
- },
258
- tflite: {
259
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolov9/tflite/camstack-yolov9c_float32.tflite"),
260
- sizeMB: 97,
261
- runtimes: ["python"]
262
- }
263
- }
264
- },
265
- // ── YOLO11 ────────────────────────────────────────────────────
266
- {
267
- id: "yolo11n",
268
- name: "YOLO11 Nano",
269
- description: "YOLO11 Nano \u2014 fastest, smallest YOLO11 detection model (mAP 39.5)",
270
- inputSize: { width: 640, height: 640 },
271
- labels: types_1.COCO_80_LABELS,
272
- formats: {
273
- onnx: {
274
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/onnx/camstack-yolo11n.onnx"),
275
- sizeMB: 10
276
- },
277
- coreml: {
278
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/coreml/camstack-yolo11n.mlpackage"),
279
- sizeMB: 5,
280
- isDirectory: true,
281
- files: exports.MLPACKAGE_FILES,
282
- runtimes: ["python"]
283
- },
284
- openvino: {
285
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/openvino/camstack-yolo11n.xml"),
286
- sizeMB: 5,
287
- runtimes: ["python"]
288
- },
289
- tflite: {
290
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/tflite/camstack-yolo11n_float32.tflite"),
291
- sizeMB: 10,
292
- runtimes: ["python"]
293
- }
294
- }
295
- },
296
- {
297
- id: "yolo11s",
298
- name: "YOLO11 Small",
299
- description: "YOLO11 Small \u2014 balanced speed and accuracy (mAP 47.0)",
300
- inputSize: { width: 640, height: 640 },
301
- labels: types_1.COCO_80_LABELS,
302
- formats: {
303
- onnx: {
304
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/onnx/camstack-yolo11s.onnx"),
305
- sizeMB: 36
306
- },
307
- coreml: {
308
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/coreml/camstack-yolo11s.mlpackage"),
309
- sizeMB: 18,
310
- isDirectory: true,
311
- files: exports.MLPACKAGE_FILES,
312
- runtimes: ["python"]
313
- },
314
- openvino: {
315
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/openvino/camstack-yolo11s.xml"),
316
- sizeMB: 18,
317
- runtimes: ["python"]
318
- },
319
- tflite: {
320
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/tflite/camstack-yolo11s_float32.tflite"),
321
- sizeMB: 36,
322
- runtimes: ["python"]
323
- }
324
- }
325
- },
326
- {
327
- id: "yolo11m",
328
- name: "YOLO11 Medium",
329
- description: "YOLO11 Medium \u2014 higher accuracy, moderate size (mAP 51.5)",
330
- inputSize: { width: 640, height: 640 },
331
- labels: types_1.COCO_80_LABELS,
332
- formats: {
333
- onnx: {
334
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/onnx/camstack-yolo11m.onnx"),
335
- sizeMB: 77
336
- },
337
- coreml: {
338
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/coreml/camstack-yolo11m.mlpackage"),
339
- sizeMB: 39,
340
- isDirectory: true,
341
- files: exports.MLPACKAGE_FILES,
342
- runtimes: ["python"]
343
- },
344
- openvino: {
345
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/openvino/camstack-yolo11m.xml"),
346
- sizeMB: 39,
347
- runtimes: ["python"]
348
- },
349
- tflite: {
350
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/tflite/camstack-yolo11m_float32.tflite"),
351
- sizeMB: 77,
352
- runtimes: ["python"]
353
- }
354
- }
355
- },
356
- {
357
- id: "yolo11l",
358
- name: "YOLO11 Large",
359
- description: "YOLO11 Large \u2014 high-accuracy large model (mAP 53.4)",
360
- inputSize: { width: 640, height: 640 },
361
- labels: types_1.COCO_80_LABELS,
362
- formats: {
363
- onnx: {
364
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/onnx/camstack-yolo11l.onnx"),
365
- sizeMB: 97
366
- },
367
- coreml: {
368
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/coreml/camstack-yolo11l.mlpackage"),
369
- sizeMB: 49,
370
- isDirectory: true,
371
- files: exports.MLPACKAGE_FILES,
372
- runtimes: ["python"]
373
- },
374
- openvino: {
375
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/openvino/camstack-yolo11l.xml"),
376
- sizeMB: 49,
377
- runtimes: ["python"]
378
- },
379
- tflite: {
380
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/tflite/camstack-yolo11l_float32.tflite"),
381
- sizeMB: 97,
382
- runtimes: ["python"]
383
- }
384
- }
385
- },
386
- {
387
- id: "yolo11x",
388
- name: "YOLO11 Extra-Large",
389
- description: "YOLO11 Extra-Large \u2014 maximum accuracy (mAP 54.7)",
390
- inputSize: { width: 640, height: 640 },
391
- labels: types_1.COCO_80_LABELS,
392
- formats: {
393
- onnx: {
394
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/onnx/camstack-yolo11x.onnx"),
395
- sizeMB: 218
396
- },
397
- coreml: {
398
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/coreml/camstack-yolo11x.mlpackage"),
399
- sizeMB: 109,
400
- isDirectory: true,
401
- files: exports.MLPACKAGE_FILES,
402
- runtimes: ["python"]
403
- },
404
- openvino: {
405
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/openvino/camstack-yolo11x.xml"),
406
- sizeMB: 109,
407
- runtimes: ["python"]
408
- },
409
- tflite: {
410
- url: (0, types_1.hfModelUrl)(HF_REPO, "objectDetection/yolo11/tflite/camstack-yolo11x_float32.tflite"),
411
- sizeMB: 218,
412
- runtimes: ["python"]
413
- }
414
- }
415
- }
416
- ];
417
- }
418
- });
419
-
420
- export {
421
- require_object_detection_models
422
- };
423
- //# sourceMappingURL=chunk-MGT6RUVX.mjs.map