@digipair/skill-tensorflow 0.25.4 → 0.25.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/index.cjs.js +7 -4
- package/index.esm.js +7 -4
- package/package.json +1 -1
- package/schema.json +18 -0
package/index.cjs.js
CHANGED
@@ -35,20 +35,23 @@ let TensorflowService = class TensorflowService {
|
|
35
35
|
return buffer;
|
36
36
|
}
|
37
37
|
async cocoSsd(params, _pinsSettingsList, _context) {
|
38
|
-
const { image } = params;
|
39
|
-
const model = await cocoSsdModel__namespace.load(
|
38
|
+
const { image, base = 'lite_mobilenet_v2', modelUrl } = params;
|
39
|
+
const model = await cocoSsdModel__namespace.load({
|
40
|
+
base,
|
41
|
+
modelUrl
|
42
|
+
});
|
40
43
|
const imageLoaded = await this.base64ToImage(image);
|
41
44
|
const imageData = tf__namespace.node.decodeImage(imageLoaded, 3);
|
42
45
|
const predictions = await model.detect(imageData);
|
43
46
|
return predictions;
|
44
47
|
}
|
45
48
|
async faceDetection(params, _pinsSettingsList, _context) {
|
46
|
-
const { image } = params;
|
49
|
+
const { image, runtime = 'mediapipe' } = params;
|
47
50
|
const model = faceDetectionModel__namespace.SupportedModels.MediaPipeFaceDetector;
|
48
51
|
const imageLoaded = await this.base64ToImage(image);
|
49
52
|
const imageData = tf__namespace.node.decodeImage(imageLoaded, 3);
|
50
53
|
const detectorConfig = {
|
51
|
-
runtime
|
54
|
+
runtime
|
52
55
|
};
|
53
56
|
const detector = await faceDetectionModel__namespace.createDetector(model, detectorConfig);
|
54
57
|
const faces = await detector.estimateFaces(imageData);
|
package/index.esm.js
CHANGED
@@ -9,20 +9,23 @@ let TensorflowService = class TensorflowService {
|
|
9
9
|
return buffer;
|
10
10
|
}
|
11
11
|
async cocoSsd(params, _pinsSettingsList, _context) {
|
12
|
-
const { image } = params;
|
13
|
-
const model = await cocoSsdModel.load(
|
12
|
+
const { image, base = 'lite_mobilenet_v2', modelUrl } = params;
|
13
|
+
const model = await cocoSsdModel.load({
|
14
|
+
base,
|
15
|
+
modelUrl
|
16
|
+
});
|
14
17
|
const imageLoaded = await this.base64ToImage(image);
|
15
18
|
const imageData = tf.node.decodeImage(imageLoaded, 3);
|
16
19
|
const predictions = await model.detect(imageData);
|
17
20
|
return predictions;
|
18
21
|
}
|
19
22
|
async faceDetection(params, _pinsSettingsList, _context) {
|
20
|
-
const { image } = params;
|
23
|
+
const { image, runtime = 'mediapipe' } = params;
|
21
24
|
const model = faceDetectionModel.SupportedModels.MediaPipeFaceDetector;
|
22
25
|
const imageLoaded = await this.base64ToImage(image);
|
23
26
|
const imageData = tf.node.decodeImage(imageLoaded, 3);
|
24
27
|
const detectorConfig = {
|
25
|
-
runtime
|
28
|
+
runtime
|
26
29
|
};
|
27
30
|
const detector = await faceDetectionModel.createDetector(model, detectorConfig);
|
28
31
|
const faces = await detector.estimateFaces(imageData);
|
package/package.json
CHANGED
package/schema.json
CHANGED
@@ -22,6 +22,15 @@
|
|
22
22
|
"schema": {
|
23
23
|
"type": "string"
|
24
24
|
}
|
25
|
+
},
|
26
|
+
{
|
27
|
+
"name": "base",
|
28
|
+
"summary": "Base",
|
29
|
+
"required": false,
|
30
|
+
"description": "Model de base (mobilenet_v1, mobilenet_v2 ou lite_mobilenet_v2)",
|
31
|
+
"schema": {
|
32
|
+
"type": "string"
|
33
|
+
}
|
25
34
|
}
|
26
35
|
],
|
27
36
|
"x-events": []
|
@@ -41,6 +50,15 @@
|
|
41
50
|
"schema": {
|
42
51
|
"type": "string"
|
43
52
|
}
|
53
|
+
},
|
54
|
+
{
|
55
|
+
"name": "runtime",
|
56
|
+
"summary": "Environnement d'exécution",
|
57
|
+
"required": false,
|
58
|
+
"description": "Environnement d'exécution (mediapipe or tfjs)",
|
59
|
+
"schema": {
|
60
|
+
"type": "string"
|
61
|
+
}
|
44
62
|
}
|
45
63
|
],
|
46
64
|
"x-events": []
|