@workglow/ai-provider 0.0.90 → 0.0.91

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (96) hide show
  1. package/README.md +45 -45
  2. package/dist/anthropic/AnthropicProvider.d.ts +39 -0
  3. package/dist/anthropic/AnthropicProvider.d.ts.map +1 -0
  4. package/dist/anthropic/Anthropic_Worker.d.ts +7 -0
  5. package/dist/anthropic/Anthropic_Worker.d.ts.map +1 -0
  6. package/dist/anthropic/common/Anthropic_Constants.d.ts +7 -0
  7. package/dist/anthropic/common/Anthropic_Constants.d.ts.map +1 -0
  8. package/dist/anthropic/common/Anthropic_JobRunFns.d.ts +16 -0
  9. package/dist/anthropic/common/Anthropic_JobRunFns.d.ts.map +1 -0
  10. package/dist/anthropic/common/Anthropic_ModelSchema.d.ts +167 -0
  11. package/dist/anthropic/common/Anthropic_ModelSchema.d.ts.map +1 -0
  12. package/dist/anthropic/index.d.ts +11 -0
  13. package/dist/anthropic/index.d.ts.map +1 -0
  14. package/dist/anthropic/index.js +257 -0
  15. package/dist/anthropic/index.js.map +14 -0
  16. package/dist/google-gemini/Gemini_Worker.d.ts +7 -0
  17. package/dist/google-gemini/Gemini_Worker.d.ts.map +1 -0
  18. package/dist/google-gemini/GoogleGeminiProvider.d.ts +36 -0
  19. package/dist/google-gemini/GoogleGeminiProvider.d.ts.map +1 -0
  20. package/dist/google-gemini/common/Gemini_Constants.d.ts +7 -0
  21. package/dist/google-gemini/common/Gemini_Constants.d.ts.map +1 -0
  22. package/dist/google-gemini/common/Gemini_JobRunFns.d.ts +17 -0
  23. package/dist/google-gemini/common/Gemini_JobRunFns.d.ts.map +1 -0
  24. package/dist/google-gemini/common/Gemini_ModelSchema.d.ts +164 -0
  25. package/dist/google-gemini/common/Gemini_ModelSchema.d.ts.map +1 -0
  26. package/dist/google-gemini/index.d.ts +11 -0
  27. package/dist/google-gemini/index.d.ts.map +1 -0
  28. package/dist/google-gemini/index.js +300 -0
  29. package/dist/google-gemini/index.js.map +14 -0
  30. package/dist/hf-transformers/HFT_Worker.d.ts +7 -0
  31. package/dist/hf-transformers/HFT_Worker.d.ts.map +1 -0
  32. package/dist/hf-transformers/HuggingFaceTransformersProvider.d.ts +41 -0
  33. package/dist/hf-transformers/HuggingFaceTransformersProvider.d.ts.map +1 -0
  34. package/dist/hf-transformers/common/HFT_JobRunFns.d.ts +1218 -1
  35. package/dist/hf-transformers/common/HFT_JobRunFns.d.ts.map +1 -1
  36. package/dist/hf-transformers/index.d.ts +2 -3
  37. package/dist/hf-transformers/index.d.ts.map +1 -1
  38. package/dist/hf-transformers/index.js +1130 -0
  39. package/dist/hf-transformers/index.js.map +15 -0
  40. package/dist/index.d.ts +23 -2
  41. package/dist/index.d.ts.map +1 -1
  42. package/dist/index.js +729 -865
  43. package/dist/index.js.map +20 -12
  44. package/dist/provider-ollama/OllamaProvider.d.ts +39 -0
  45. package/dist/provider-ollama/OllamaProvider.d.ts.map +1 -0
  46. package/dist/provider-ollama/Ollama_Worker.d.ts +7 -0
  47. package/dist/provider-ollama/Ollama_Worker.d.ts.map +1 -0
  48. package/dist/provider-ollama/common/Ollama_Constants.d.ts +8 -0
  49. package/dist/provider-ollama/common/Ollama_Constants.d.ts.map +1 -0
  50. package/dist/provider-ollama/common/Ollama_JobRunFns.d.ts +17 -0
  51. package/dist/provider-ollama/common/Ollama_JobRunFns.d.ts.map +1 -0
  52. package/dist/provider-ollama/common/Ollama_ModelSchema.d.ts +137 -0
  53. package/dist/provider-ollama/common/Ollama_ModelSchema.d.ts.map +1 -0
  54. package/dist/provider-ollama/index.d.ts +11 -0
  55. package/dist/provider-ollama/index.d.ts.map +1 -0
  56. package/dist/provider-ollama/index.js +274 -0
  57. package/dist/provider-ollama/index.js.map +14 -0
  58. package/dist/provider-openai/OpenAI_Worker.d.ts +7 -0
  59. package/dist/provider-openai/OpenAI_Worker.d.ts.map +1 -0
  60. package/dist/provider-openai/OpenAiProvider.d.ts +40 -0
  61. package/dist/provider-openai/OpenAiProvider.d.ts.map +1 -0
  62. package/dist/provider-openai/common/OpenAI_Constants.d.ts +7 -0
  63. package/dist/provider-openai/common/OpenAI_Constants.d.ts.map +1 -0
  64. package/dist/provider-openai/common/OpenAI_JobRunFns.d.ts +17 -0
  65. package/dist/provider-openai/common/OpenAI_JobRunFns.d.ts.map +1 -0
  66. package/dist/provider-openai/common/OpenAI_ModelSchema.d.ts +161 -0
  67. package/dist/provider-openai/common/OpenAI_ModelSchema.d.ts.map +1 -0
  68. package/dist/provider-openai/index.d.ts +11 -0
  69. package/dist/provider-openai/index.d.ts.map +1 -0
  70. package/dist/provider-openai/index.js +283 -0
  71. package/dist/provider-openai/index.js.map +14 -0
  72. package/dist/tf-mediapipe/TFMP_Worker.d.ts +7 -0
  73. package/dist/tf-mediapipe/TFMP_Worker.d.ts.map +1 -0
  74. package/dist/tf-mediapipe/TensorFlowMediaPipeProvider.d.ts +39 -0
  75. package/dist/tf-mediapipe/TensorFlowMediaPipeProvider.d.ts.map +1 -0
  76. package/dist/tf-mediapipe/common/TFMP_JobRunFns.d.ts +842 -0
  77. package/dist/tf-mediapipe/common/TFMP_JobRunFns.d.ts.map +1 -1
  78. package/dist/tf-mediapipe/index.d.ts +2 -3
  79. package/dist/tf-mediapipe/index.d.ts.map +1 -1
  80. package/dist/tf-mediapipe/index.js +674 -0
  81. package/dist/tf-mediapipe/index.js.map +14 -0
  82. package/package.json +83 -18
  83. package/dist/hf-transformers/registry/HFT_Client_RegisterJobFns.d.ts +0 -17
  84. package/dist/hf-transformers/registry/HFT_Client_RegisterJobFns.d.ts.map +0 -1
  85. package/dist/hf-transformers/registry/HFT_Inline_RegisterJobFns.d.ts +0 -16
  86. package/dist/hf-transformers/registry/HFT_Inline_RegisterJobFns.d.ts.map +0 -1
  87. package/dist/hf-transformers/registry/HFT_Worker_RegisterJobFns.d.ts +0 -8
  88. package/dist/hf-transformers/registry/HFT_Worker_RegisterJobFns.d.ts.map +0 -1
  89. package/dist/openai/provider/OpenAI_TaskRun.d.ts +0 -2
  90. package/dist/openai/provider/OpenAI_TaskRun.d.ts.map +0 -1
  91. package/dist/tf-mediapipe/registry/TFMP_Client_RegisterJobFns.d.ts +0 -17
  92. package/dist/tf-mediapipe/registry/TFMP_Client_RegisterJobFns.d.ts.map +0 -1
  93. package/dist/tf-mediapipe/registry/TFMP_Inline_RegisterJobFns.d.ts +0 -16
  94. package/dist/tf-mediapipe/registry/TFMP_Inline_RegisterJobFns.d.ts.map +0 -1
  95. package/dist/tf-mediapipe/registry/TFMP_Worker_RegisterJobFns.d.ts +0 -8
  96. package/dist/tf-mediapipe/registry/TFMP_Worker_RegisterJobFns.d.ts.map +0 -1
@@ -71,4 +71,846 @@ export declare const TFMP_FaceLandmarker: AiProviderRunFn<FaceLandmarkerTaskInpu
71
71
  * Core implementation for pose landmark detection using MediaPipe.
72
72
  */
73
73
  export declare const TFMP_PoseLandmarker: AiProviderRunFn<PoseLandmarkerTaskInput, PoseLandmarkerTaskOutput, TFMPModelConfig>;
74
+ /**
75
+ * All TensorFlow MediaPipe task run functions, keyed by task type name.
76
+ * Pass this to `new TensorFlowMediaPipeProvider(TFMP_TASKS)` when the
77
+ * actual run function implementations are needed (inline mode, worker server).
78
+ */
79
+ export declare const TFMP_TASKS: {
80
+ readonly DownloadModelTask: AiProviderRunFn<{
81
+ model: string | {
82
+ [x: string]: unknown;
83
+ title?: string | undefined;
84
+ description?: string | undefined;
85
+ model_id?: string | undefined;
86
+ tasks?: string[] | undefined;
87
+ metadata?: {
88
+ [x: string]: unknown;
89
+ } | undefined;
90
+ provider: string;
91
+ provider_config: {
92
+ [x: string]: unknown;
93
+ };
94
+ };
95
+ }, {
96
+ model: string | {
97
+ [x: string]: unknown;
98
+ title?: string | undefined;
99
+ description?: string | undefined;
100
+ model_id?: string | undefined;
101
+ tasks?: string[] | undefined;
102
+ metadata?: {
103
+ [x: string]: unknown;
104
+ } | undefined;
105
+ provider: string;
106
+ provider_config: {
107
+ [x: string]: unknown;
108
+ };
109
+ };
110
+ }, {
111
+ title?: string | undefined;
112
+ description?: string | undefined;
113
+ model_id?: string | undefined;
114
+ tasks?: string[] | undefined;
115
+ metadata?: {
116
+ [x: string]: unknown;
117
+ } | undefined;
118
+ provider: "TENSORFLOW_MEDIAPIPE";
119
+ provider_config: {
120
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
121
+ model_path: string;
122
+ task_engine: "text" | "audio" | "vision" | "genai";
123
+ };
124
+ }>;
125
+ readonly UnloadModelTask: AiProviderRunFn<{
126
+ model: string | {
127
+ [x: string]: unknown;
128
+ title?: string | undefined;
129
+ description?: string | undefined;
130
+ model_id?: string | undefined;
131
+ tasks?: string[] | undefined;
132
+ metadata?: {
133
+ [x: string]: unknown;
134
+ } | undefined;
135
+ provider: string;
136
+ provider_config: {
137
+ [x: string]: unknown;
138
+ };
139
+ };
140
+ }, {
141
+ model: string | {
142
+ [x: string]: unknown;
143
+ title?: string | undefined;
144
+ description?: string | undefined;
145
+ model_id?: string | undefined;
146
+ tasks?: string[] | undefined;
147
+ metadata?: {
148
+ [x: string]: unknown;
149
+ } | undefined;
150
+ provider: string;
151
+ provider_config: {
152
+ [x: string]: unknown;
153
+ };
154
+ };
155
+ }, {
156
+ title?: string | undefined;
157
+ description?: string | undefined;
158
+ model_id?: string | undefined;
159
+ tasks?: string[] | undefined;
160
+ metadata?: {
161
+ [x: string]: unknown;
162
+ } | undefined;
163
+ provider: "TENSORFLOW_MEDIAPIPE";
164
+ provider_config: {
165
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
166
+ model_path: string;
167
+ task_engine: "text" | "audio" | "vision" | "genai";
168
+ };
169
+ }>;
170
+ readonly TextEmbeddingTask: AiProviderRunFn<{
171
+ model: string | {
172
+ [x: string]: unknown;
173
+ title?: string | undefined;
174
+ description?: string | undefined;
175
+ model_id?: string | undefined;
176
+ tasks?: string[] | undefined;
177
+ metadata?: {
178
+ [x: string]: unknown;
179
+ } | undefined;
180
+ provider: string;
181
+ provider_config: {
182
+ [x: string]: unknown;
183
+ };
184
+ };
185
+ text: string | string[];
186
+ }, {
187
+ vector: import("@workglow/util").TypedArray | import("@workglow/util").TypedArray[];
188
+ }, {
189
+ title?: string | undefined;
190
+ description?: string | undefined;
191
+ model_id?: string | undefined;
192
+ tasks?: string[] | undefined;
193
+ metadata?: {
194
+ [x: string]: unknown;
195
+ } | undefined;
196
+ provider: "TENSORFLOW_MEDIAPIPE";
197
+ provider_config: {
198
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
199
+ model_path: string;
200
+ task_engine: "text" | "audio" | "vision" | "genai";
201
+ };
202
+ }>;
203
+ readonly TextLanguageDetectionTask: AiProviderRunFn<{
204
+ maxLanguages?: number | undefined;
205
+ model: string | {
206
+ [x: string]: unknown;
207
+ title?: string | undefined;
208
+ description?: string | undefined;
209
+ model_id?: string | undefined;
210
+ tasks?: string[] | undefined;
211
+ metadata?: {
212
+ [x: string]: unknown;
213
+ } | undefined;
214
+ provider: string;
215
+ provider_config: {
216
+ [x: string]: unknown;
217
+ };
218
+ };
219
+ text: string;
220
+ }, {
221
+ languages: {
222
+ score: number;
223
+ language: string;
224
+ }[];
225
+ }, {
226
+ title?: string | undefined;
227
+ description?: string | undefined;
228
+ model_id?: string | undefined;
229
+ tasks?: string[] | undefined;
230
+ metadata?: {
231
+ [x: string]: unknown;
232
+ } | undefined;
233
+ provider: "TENSORFLOW_MEDIAPIPE";
234
+ provider_config: {
235
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
236
+ model_path: string;
237
+ task_engine: "text" | "audio" | "vision" | "genai";
238
+ };
239
+ }>;
240
+ readonly TextClassificationTask: AiProviderRunFn<{
241
+ candidateLabels?: string[] | undefined;
242
+ maxCategories?: number | undefined;
243
+ model: string | {
244
+ [x: string]: unknown;
245
+ title?: string | undefined;
246
+ description?: string | undefined;
247
+ model_id?: string | undefined;
248
+ tasks?: string[] | undefined;
249
+ metadata?: {
250
+ [x: string]: unknown;
251
+ } | undefined;
252
+ provider: string;
253
+ provider_config: {
254
+ [x: string]: unknown;
255
+ };
256
+ };
257
+ text: string;
258
+ }, {
259
+ categories: {
260
+ label: string;
261
+ score: number;
262
+ }[];
263
+ }, {
264
+ title?: string | undefined;
265
+ description?: string | undefined;
266
+ model_id?: string | undefined;
267
+ tasks?: string[] | undefined;
268
+ metadata?: {
269
+ [x: string]: unknown;
270
+ } | undefined;
271
+ provider: "TENSORFLOW_MEDIAPIPE";
272
+ provider_config: {
273
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
274
+ model_path: string;
275
+ task_engine: "text" | "audio" | "vision" | "genai";
276
+ };
277
+ }>;
278
+ readonly ImageSegmentationTask: AiProviderRunFn<{
279
+ threshold?: number | undefined;
280
+ maskThreshold?: number | undefined;
281
+ model: string | {
282
+ [x: string]: unknown;
283
+ title?: string | undefined;
284
+ description?: string | undefined;
285
+ model_id?: string | undefined;
286
+ tasks?: string[] | undefined;
287
+ metadata?: {
288
+ [x: string]: unknown;
289
+ } | undefined;
290
+ provider: string;
291
+ provider_config: {
292
+ [x: string]: unknown;
293
+ };
294
+ };
295
+ image: string | {
296
+ data: {
297
+ [x: string]: unknown;
298
+ } | {
299
+ rawChannels?: number | undefined;
300
+ data: number[];
301
+ width: number;
302
+ height: number;
303
+ channels: number;
304
+ };
305
+ width: number;
306
+ height: number;
307
+ channels: number;
308
+ };
309
+ }, {
310
+ masks: {
311
+ label: string;
312
+ score: number;
313
+ mask: {
314
+ [x: string]: unknown;
315
+ };
316
+ }[] | {
317
+ label: string;
318
+ score: number;
319
+ mask: {
320
+ [x: string]: unknown;
321
+ };
322
+ }[][];
323
+ }, {
324
+ title?: string | undefined;
325
+ description?: string | undefined;
326
+ model_id?: string | undefined;
327
+ tasks?: string[] | undefined;
328
+ metadata?: {
329
+ [x: string]: unknown;
330
+ } | undefined;
331
+ provider: "TENSORFLOW_MEDIAPIPE";
332
+ provider_config: {
333
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
334
+ model_path: string;
335
+ task_engine: "text" | "audio" | "vision" | "genai";
336
+ };
337
+ }>;
338
+ readonly ImageEmbeddingTask: AiProviderRunFn<{
339
+ model: string | {
340
+ [x: string]: unknown;
341
+ title?: string | undefined;
342
+ description?: string | undefined;
343
+ model_id?: string | undefined;
344
+ tasks?: string[] | undefined;
345
+ metadata?: {
346
+ [x: string]: unknown;
347
+ } | undefined;
348
+ provider: string;
349
+ provider_config: {
350
+ [x: string]: unknown;
351
+ };
352
+ };
353
+ image: string | {
354
+ data: {
355
+ [x: string]: unknown;
356
+ } | {
357
+ rawChannels?: number | undefined;
358
+ data: number[];
359
+ width: number;
360
+ height: number;
361
+ channels: number;
362
+ };
363
+ width: number;
364
+ height: number;
365
+ channels: number;
366
+ };
367
+ }, {
368
+ vector: import("@workglow/util").TypedArray;
369
+ }, {
370
+ title?: string | undefined;
371
+ description?: string | undefined;
372
+ model_id?: string | undefined;
373
+ tasks?: string[] | undefined;
374
+ metadata?: {
375
+ [x: string]: unknown;
376
+ } | undefined;
377
+ provider: "TENSORFLOW_MEDIAPIPE";
378
+ provider_config: {
379
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
380
+ model_path: string;
381
+ task_engine: "text" | "audio" | "vision" | "genai";
382
+ };
383
+ }>;
384
+ readonly ImageClassificationTask: AiProviderRunFn<{
385
+ maxCategories?: number | undefined;
386
+ categories?: string[] | undefined;
387
+ model: string | {
388
+ [x: string]: unknown;
389
+ title?: string | undefined;
390
+ description?: string | undefined;
391
+ model_id?: string | undefined;
392
+ tasks?: string[] | undefined;
393
+ metadata?: {
394
+ [x: string]: unknown;
395
+ } | undefined;
396
+ provider: string;
397
+ provider_config: {
398
+ [x: string]: unknown;
399
+ };
400
+ };
401
+ image: string | {
402
+ data: {
403
+ [x: string]: unknown;
404
+ } | {
405
+ rawChannels?: number | undefined;
406
+ data: number[];
407
+ width: number;
408
+ height: number;
409
+ channels: number;
410
+ };
411
+ width: number;
412
+ height: number;
413
+ channels: number;
414
+ };
415
+ }, {
416
+ categories: {
417
+ label: string;
418
+ score: number;
419
+ }[] | {
420
+ label: string;
421
+ score: number;
422
+ }[][];
423
+ }, {
424
+ title?: string | undefined;
425
+ description?: string | undefined;
426
+ model_id?: string | undefined;
427
+ tasks?: string[] | undefined;
428
+ metadata?: {
429
+ [x: string]: unknown;
430
+ } | undefined;
431
+ provider: "TENSORFLOW_MEDIAPIPE";
432
+ provider_config: {
433
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
434
+ model_path: string;
435
+ task_engine: "text" | "audio" | "vision" | "genai";
436
+ };
437
+ }>;
438
+ readonly ObjectDetectionTask: AiProviderRunFn<{
439
+ threshold?: number | undefined;
440
+ labels?: string[] | undefined;
441
+ model: string | {
442
+ [x: string]: unknown;
443
+ title?: string | undefined;
444
+ description?: string | undefined;
445
+ model_id?: string | undefined;
446
+ tasks?: string[] | undefined;
447
+ metadata?: {
448
+ [x: string]: unknown;
449
+ } | undefined;
450
+ provider: string;
451
+ provider_config: {
452
+ [x: string]: unknown;
453
+ };
454
+ };
455
+ image: string | {
456
+ data: {
457
+ [x: string]: unknown;
458
+ } | {
459
+ rawChannels?: number | undefined;
460
+ data: number[];
461
+ width: number;
462
+ height: number;
463
+ channels: number;
464
+ };
465
+ width: number;
466
+ height: number;
467
+ channels: number;
468
+ };
469
+ }, {
470
+ detections: {
471
+ label: string;
472
+ score: number;
473
+ box: {
474
+ width: number;
475
+ height: number;
476
+ x: number;
477
+ y: number;
478
+ };
479
+ }[] | {
480
+ label: string;
481
+ score: number;
482
+ box: {
483
+ width: number;
484
+ height: number;
485
+ x: number;
486
+ y: number;
487
+ };
488
+ }[][];
489
+ }, {
490
+ title?: string | undefined;
491
+ description?: string | undefined;
492
+ model_id?: string | undefined;
493
+ tasks?: string[] | undefined;
494
+ metadata?: {
495
+ [x: string]: unknown;
496
+ } | undefined;
497
+ provider: "TENSORFLOW_MEDIAPIPE";
498
+ provider_config: {
499
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
500
+ model_path: string;
501
+ task_engine: "text" | "audio" | "vision" | "genai";
502
+ };
503
+ }>;
504
+ readonly GestureRecognizerTask: AiProviderRunFn<{
505
+ numHands?: number | undefined;
506
+ minHandDetectionConfidence?: number | undefined;
507
+ minHandPresenceConfidence?: number | undefined;
508
+ minTrackingConfidence?: number | undefined;
509
+ model: string | {
510
+ [x: string]: unknown;
511
+ title?: string | undefined;
512
+ description?: string | undefined;
513
+ model_id?: string | undefined;
514
+ tasks?: string[] | undefined;
515
+ metadata?: {
516
+ [x: string]: unknown;
517
+ } | undefined;
518
+ provider: string;
519
+ provider_config: {
520
+ [x: string]: unknown;
521
+ };
522
+ };
523
+ image: string | {
524
+ data: {
525
+ [x: string]: unknown;
526
+ } | {
527
+ rawChannels?: number | undefined;
528
+ data: number[];
529
+ width: number;
530
+ height: number;
531
+ channels: number;
532
+ };
533
+ width: number;
534
+ height: number;
535
+ channels: number;
536
+ };
537
+ }, {
538
+ hands: {
539
+ gestures: {
540
+ label: string;
541
+ score: number;
542
+ }[];
543
+ handedness: {
544
+ label: string;
545
+ score: number;
546
+ }[];
547
+ landmarks: {
548
+ x: number;
549
+ y: number;
550
+ z: number;
551
+ }[];
552
+ worldLandmarks: {
553
+ x: number;
554
+ y: number;
555
+ z: number;
556
+ }[];
557
+ }[] | {
558
+ gestures: {
559
+ label: string;
560
+ score: number;
561
+ }[];
562
+ handedness: {
563
+ label: string;
564
+ score: number;
565
+ }[];
566
+ landmarks: {
567
+ x: number;
568
+ y: number;
569
+ z: number;
570
+ }[];
571
+ worldLandmarks: {
572
+ x: number;
573
+ y: number;
574
+ z: number;
575
+ }[];
576
+ }[][];
577
+ }, {
578
+ title?: string | undefined;
579
+ description?: string | undefined;
580
+ model_id?: string | undefined;
581
+ tasks?: string[] | undefined;
582
+ metadata?: {
583
+ [x: string]: unknown;
584
+ } | undefined;
585
+ provider: "TENSORFLOW_MEDIAPIPE";
586
+ provider_config: {
587
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
588
+ model_path: string;
589
+ task_engine: "text" | "audio" | "vision" | "genai";
590
+ };
591
+ }>;
592
+ readonly HandLandmarkerTask: AiProviderRunFn<{
593
+ numHands?: number | undefined;
594
+ minHandDetectionConfidence?: number | undefined;
595
+ minHandPresenceConfidence?: number | undefined;
596
+ minTrackingConfidence?: number | undefined;
597
+ model: string | {
598
+ [x: string]: unknown;
599
+ title?: string | undefined;
600
+ description?: string | undefined;
601
+ model_id?: string | undefined;
602
+ tasks?: string[] | undefined;
603
+ metadata?: {
604
+ [x: string]: unknown;
605
+ } | undefined;
606
+ provider: string;
607
+ provider_config: {
608
+ [x: string]: unknown;
609
+ };
610
+ };
611
+ image: string | {
612
+ data: {
613
+ [x: string]: unknown;
614
+ } | {
615
+ rawChannels?: number | undefined;
616
+ data: number[];
617
+ width: number;
618
+ height: number;
619
+ channels: number;
620
+ };
621
+ width: number;
622
+ height: number;
623
+ channels: number;
624
+ };
625
+ }, {
626
+ hands: {
627
+ handedness: {
628
+ label: string;
629
+ score: number;
630
+ }[];
631
+ landmarks: {
632
+ x: number;
633
+ y: number;
634
+ z: number;
635
+ }[];
636
+ worldLandmarks: {
637
+ x: number;
638
+ y: number;
639
+ z: number;
640
+ }[];
641
+ }[] | {
642
+ handedness: {
643
+ label: string;
644
+ score: number;
645
+ }[];
646
+ landmarks: {
647
+ x: number;
648
+ y: number;
649
+ z: number;
650
+ }[];
651
+ worldLandmarks: {
652
+ x: number;
653
+ y: number;
654
+ z: number;
655
+ }[];
656
+ }[][];
657
+ }, {
658
+ title?: string | undefined;
659
+ description?: string | undefined;
660
+ model_id?: string | undefined;
661
+ tasks?: string[] | undefined;
662
+ metadata?: {
663
+ [x: string]: unknown;
664
+ } | undefined;
665
+ provider: "TENSORFLOW_MEDIAPIPE";
666
+ provider_config: {
667
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
668
+ model_path: string;
669
+ task_engine: "text" | "audio" | "vision" | "genai";
670
+ };
671
+ }>;
672
+ readonly FaceDetectorTask: AiProviderRunFn<{
673
+ minDetectionConfidence?: number | undefined;
674
+ minSuppressionThreshold?: number | undefined;
675
+ model: string | {
676
+ [x: string]: unknown;
677
+ title?: string | undefined;
678
+ description?: string | undefined;
679
+ model_id?: string | undefined;
680
+ tasks?: string[] | undefined;
681
+ metadata?: {
682
+ [x: string]: unknown;
683
+ } | undefined;
684
+ provider: string;
685
+ provider_config: {
686
+ [x: string]: unknown;
687
+ };
688
+ };
689
+ image: string | {
690
+ data: {
691
+ [x: string]: unknown;
692
+ } | {
693
+ rawChannels?: number | undefined;
694
+ data: number[];
695
+ width: number;
696
+ height: number;
697
+ channels: number;
698
+ };
699
+ width: number;
700
+ height: number;
701
+ channels: number;
702
+ };
703
+ }, {
704
+ faces: {
705
+ score: number;
706
+ box: {
707
+ width: number;
708
+ height: number;
709
+ x: number;
710
+ y: number;
711
+ };
712
+ keypoints: {
713
+ label?: string | undefined;
714
+ x: number;
715
+ y: number;
716
+ }[];
717
+ }[] | {
718
+ score: number;
719
+ box: {
720
+ width: number;
721
+ height: number;
722
+ x: number;
723
+ y: number;
724
+ };
725
+ keypoints: {
726
+ label?: string | undefined;
727
+ x: number;
728
+ y: number;
729
+ }[];
730
+ }[][];
731
+ }, {
732
+ title?: string | undefined;
733
+ description?: string | undefined;
734
+ model_id?: string | undefined;
735
+ tasks?: string[] | undefined;
736
+ metadata?: {
737
+ [x: string]: unknown;
738
+ } | undefined;
739
+ provider: "TENSORFLOW_MEDIAPIPE";
740
+ provider_config: {
741
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
742
+ model_path: string;
743
+ task_engine: "text" | "audio" | "vision" | "genai";
744
+ };
745
+ }>;
746
+ readonly FaceLandmarkerTask: AiProviderRunFn<{
747
+ minTrackingConfidence?: number | undefined;
748
+ numFaces?: number | undefined;
749
+ minFaceDetectionConfidence?: number | undefined;
750
+ minFacePresenceConfidence?: number | undefined;
751
+ outputFaceBlendshapes?: boolean | undefined;
752
+ outputFacialTransformationMatrixes?: boolean | undefined;
753
+ model: string | {
754
+ [x: string]: unknown;
755
+ title?: string | undefined;
756
+ description?: string | undefined;
757
+ model_id?: string | undefined;
758
+ tasks?: string[] | undefined;
759
+ metadata?: {
760
+ [x: string]: unknown;
761
+ } | undefined;
762
+ provider: string;
763
+ provider_config: {
764
+ [x: string]: unknown;
765
+ };
766
+ };
767
+ image: string | {
768
+ data: {
769
+ [x: string]: unknown;
770
+ } | {
771
+ rawChannels?: number | undefined;
772
+ data: number[];
773
+ width: number;
774
+ height: number;
775
+ channels: number;
776
+ };
777
+ width: number;
778
+ height: number;
779
+ channels: number;
780
+ };
781
+ }, {
782
+ faces: {
783
+ blendshapes?: {
784
+ label: string;
785
+ score: number;
786
+ }[] | undefined;
787
+ transformationMatrix?: number[] | undefined;
788
+ landmarks: {
789
+ x: number;
790
+ y: number;
791
+ z: number;
792
+ }[];
793
+ }[] | {
794
+ blendshapes?: {
795
+ label: string;
796
+ score: number;
797
+ }[] | undefined;
798
+ transformationMatrix?: number[] | undefined;
799
+ landmarks: {
800
+ x: number;
801
+ y: number;
802
+ z: number;
803
+ }[];
804
+ }[][];
805
+ }, {
806
+ title?: string | undefined;
807
+ description?: string | undefined;
808
+ model_id?: string | undefined;
809
+ tasks?: string[] | undefined;
810
+ metadata?: {
811
+ [x: string]: unknown;
812
+ } | undefined;
813
+ provider: "TENSORFLOW_MEDIAPIPE";
814
+ provider_config: {
815
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
816
+ model_path: string;
817
+ task_engine: "text" | "audio" | "vision" | "genai";
818
+ };
819
+ }>;
820
+ readonly PoseLandmarkerTask: AiProviderRunFn<{
821
+ minTrackingConfidence?: number | undefined;
822
+ numPoses?: number | undefined;
823
+ minPoseDetectionConfidence?: number | undefined;
824
+ minPosePresenceConfidence?: number | undefined;
825
+ outputSegmentationMasks?: boolean | undefined;
826
+ model: string | {
827
+ [x: string]: unknown;
828
+ title?: string | undefined;
829
+ description?: string | undefined;
830
+ model_id?: string | undefined;
831
+ tasks?: string[] | undefined;
832
+ metadata?: {
833
+ [x: string]: unknown;
834
+ } | undefined;
835
+ provider: string;
836
+ provider_config: {
837
+ [x: string]: unknown;
838
+ };
839
+ };
840
+ image: string | {
841
+ data: {
842
+ [x: string]: unknown;
843
+ } | {
844
+ rawChannels?: number | undefined;
845
+ data: number[];
846
+ width: number;
847
+ height: number;
848
+ channels: number;
849
+ };
850
+ width: number;
851
+ height: number;
852
+ channels: number;
853
+ };
854
+ }, {
855
+ poses: {
856
+ segmentationMask?: {
857
+ data: {
858
+ [x: string]: unknown;
859
+ };
860
+ width: number;
861
+ height: number;
862
+ } | undefined;
863
+ landmarks: {
864
+ visibility?: number | undefined;
865
+ presence?: number | undefined;
866
+ x: number;
867
+ y: number;
868
+ z: number;
869
+ }[];
870
+ worldLandmarks: {
871
+ visibility?: number | undefined;
872
+ presence?: number | undefined;
873
+ x: number;
874
+ y: number;
875
+ z: number;
876
+ }[];
877
+ }[] | {
878
+ segmentationMask?: {
879
+ data: {
880
+ [x: string]: unknown;
881
+ };
882
+ width: number;
883
+ height: number;
884
+ } | undefined;
885
+ landmarks: {
886
+ visibility?: number | undefined;
887
+ presence?: number | undefined;
888
+ x: number;
889
+ y: number;
890
+ z: number;
891
+ }[];
892
+ worldLandmarks: {
893
+ visibility?: number | undefined;
894
+ presence?: number | undefined;
895
+ x: number;
896
+ y: number;
897
+ z: number;
898
+ }[];
899
+ }[][];
900
+ }, {
901
+ title?: string | undefined;
902
+ description?: string | undefined;
903
+ model_id?: string | undefined;
904
+ tasks?: string[] | undefined;
905
+ metadata?: {
906
+ [x: string]: unknown;
907
+ } | undefined;
908
+ provider: "TENSORFLOW_MEDIAPIPE";
909
+ provider_config: {
910
+ pipeline: "text-embedder" | "text-classifier" | "text-language-detector" | "genai-text" | "audio-classifier" | "audio-embedder" | "vision-face-detector" | "vision-face-landmarker" | "vision-face-stylizer" | "vision-gesture-recognizer" | "vision-hand-landmarker" | "vision-holistic-landmarker" | "vision-image-classifier" | "vision-image-embedder" | "vision-image-segmenter" | "vision-image-interactive-segmenter" | "vision-object-detector" | "vision-pose-landmarker";
911
+ model_path: string;
912
+ task_engine: "text" | "audio" | "vision" | "genai";
913
+ };
914
+ }>;
915
+ };
74
916
  //# sourceMappingURL=TFMP_JobRunFns.d.ts.map