@digipair/skill-tensorflow 0.25.5 → 0.26.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/index.cjs.js +7 -4
- package/index.esm.js +7 -4
- package/package.json +1 -1
- package/schema.fr.json +71 -0
- package/schema.json +26 -8
package/index.cjs.js
CHANGED
@@ -35,20 +35,23 @@ let TensorflowService = class TensorflowService {
|
|
35
35
|
return buffer;
|
36
36
|
}
|
37
37
|
async cocoSsd(params, _pinsSettingsList, _context) {
|
38
|
-
const { image } = params;
|
39
|
-
const model = await cocoSsdModel__namespace.load(
|
38
|
+
const { image, base = 'lite_mobilenet_v2', modelUrl } = params;
|
39
|
+
const model = await cocoSsdModel__namespace.load({
|
40
|
+
base,
|
41
|
+
modelUrl
|
42
|
+
});
|
40
43
|
const imageLoaded = await this.base64ToImage(image);
|
41
44
|
const imageData = tf__namespace.node.decodeImage(imageLoaded, 3);
|
42
45
|
const predictions = await model.detect(imageData);
|
43
46
|
return predictions;
|
44
47
|
}
|
45
48
|
async faceDetection(params, _pinsSettingsList, _context) {
|
46
|
-
const { image } = params;
|
49
|
+
const { image, runtime = 'mediapipe' } = params;
|
47
50
|
const model = faceDetectionModel__namespace.SupportedModels.MediaPipeFaceDetector;
|
48
51
|
const imageLoaded = await this.base64ToImage(image);
|
49
52
|
const imageData = tf__namespace.node.decodeImage(imageLoaded, 3);
|
50
53
|
const detectorConfig = {
|
51
|
-
runtime
|
54
|
+
runtime
|
52
55
|
};
|
53
56
|
const detector = await faceDetectionModel__namespace.createDetector(model, detectorConfig);
|
54
57
|
const faces = await detector.estimateFaces(imageData);
|
package/index.esm.js
CHANGED
@@ -9,20 +9,23 @@ let TensorflowService = class TensorflowService {
|
|
9
9
|
return buffer;
|
10
10
|
}
|
11
11
|
async cocoSsd(params, _pinsSettingsList, _context) {
|
12
|
-
const { image } = params;
|
13
|
-
const model = await cocoSsdModel.load(
|
12
|
+
const { image, base = 'lite_mobilenet_v2', modelUrl } = params;
|
13
|
+
const model = await cocoSsdModel.load({
|
14
|
+
base,
|
15
|
+
modelUrl
|
16
|
+
});
|
14
17
|
const imageLoaded = await this.base64ToImage(image);
|
15
18
|
const imageData = tf.node.decodeImage(imageLoaded, 3);
|
16
19
|
const predictions = await model.detect(imageData);
|
17
20
|
return predictions;
|
18
21
|
}
|
19
22
|
async faceDetection(params, _pinsSettingsList, _context) {
|
20
|
-
const { image } = params;
|
23
|
+
const { image, runtime = 'mediapipe' } = params;
|
21
24
|
const model = faceDetectionModel.SupportedModels.MediaPipeFaceDetector;
|
22
25
|
const imageLoaded = await this.base64ToImage(image);
|
23
26
|
const imageData = tf.node.decodeImage(imageLoaded, 3);
|
24
27
|
const detectorConfig = {
|
25
|
-
runtime
|
28
|
+
runtime
|
26
29
|
};
|
27
30
|
const detector = await faceDetectionModel.createDetector(model, detectorConfig);
|
28
31
|
const faces = await detector.estimateFaces(imageData);
|
package/package.json
CHANGED
package/schema.fr.json
ADDED
@@ -0,0 +1,71 @@
|
|
1
|
+
{
|
2
|
+
"openapi": "3.0.0",
|
3
|
+
"info": {
|
4
|
+
"title": "@digipair/skill-tensorflow",
|
5
|
+
"summary": "Analyse par Deep Learning",
|
6
|
+
"description": "Cette compétence permet aux utilisateurs d'analyser des données par Deep Learning.",
|
7
|
+
"version": "0.1.0",
|
8
|
+
"x-icon": "🤖"
|
9
|
+
},
|
10
|
+
"paths": {
|
11
|
+
"/cocoSsd": {
|
12
|
+
"post": {
|
13
|
+
"tags": ["service"],
|
14
|
+
"summary": "Détecte les objets",
|
15
|
+
"description": "Détecte les objets dans une image",
|
16
|
+
"parameters": [
|
17
|
+
{
|
18
|
+
"name": "image",
|
19
|
+
"summary": "Image",
|
20
|
+
"required": true,
|
21
|
+
"description": "Image à analyser",
|
22
|
+
"schema": {
|
23
|
+
"type": "string"
|
24
|
+
}
|
25
|
+
},
|
26
|
+
{
|
27
|
+
"name": "base",
|
28
|
+
"summary": "Base",
|
29
|
+
"required": false,
|
30
|
+
"description": "Model de base (mobilenet_v1, mobilenet_v2 ou lite_mobilenet_v2)",
|
31
|
+
"schema": {
|
32
|
+
"type": "string"
|
33
|
+
}
|
34
|
+
}
|
35
|
+
],
|
36
|
+
"x-events": []
|
37
|
+
}
|
38
|
+
},
|
39
|
+
"/faceDetection": {
|
40
|
+
"post": {
|
41
|
+
"tags": ["service"],
|
42
|
+
"summary": "Détecte les visages",
|
43
|
+
"description": "Détecte les visages dans une image",
|
44
|
+
"parameters": [
|
45
|
+
{
|
46
|
+
"name": "image",
|
47
|
+
"summary": "Image",
|
48
|
+
"required": true,
|
49
|
+
"description": "Image à analyser",
|
50
|
+
"schema": {
|
51
|
+
"type": "string"
|
52
|
+
}
|
53
|
+
},
|
54
|
+
{
|
55
|
+
"name": "runtime",
|
56
|
+
"summary": "Environnement d'exécution",
|
57
|
+
"required": false,
|
58
|
+
"description": "Environnement d'exécution (mediapipe or tfjs)",
|
59
|
+
"schema": {
|
60
|
+
"type": "string"
|
61
|
+
}
|
62
|
+
}
|
63
|
+
],
|
64
|
+
"x-events": []
|
65
|
+
}
|
66
|
+
}
|
67
|
+
},
|
68
|
+
"components": {
|
69
|
+
"schemas": {}
|
70
|
+
}
|
71
|
+
}
|
package/schema.json
CHANGED
@@ -2,8 +2,8 @@
|
|
2
2
|
"openapi": "3.0.0",
|
3
3
|
"info": {
|
4
4
|
"title": "@digipair/skill-tensorflow",
|
5
|
-
"summary": "
|
6
|
-
"description": "
|
5
|
+
"summary": "Deep Learning Analysis",
|
6
|
+
"description": "This skill allows users to analyze data using Deep Learning.",
|
7
7
|
"version": "0.1.0",
|
8
8
|
"x-icon": "🤖"
|
9
9
|
},
|
@@ -11,14 +11,23 @@
|
|
11
11
|
"/cocoSsd": {
|
12
12
|
"post": {
|
13
13
|
"tags": ["service"],
|
14
|
-
"summary": "
|
15
|
-
"description": "
|
14
|
+
"summary": "Detects objects",
|
15
|
+
"description": "Detects objects in an image",
|
16
16
|
"parameters": [
|
17
17
|
{
|
18
18
|
"name": "image",
|
19
19
|
"summary": "Image",
|
20
20
|
"required": true,
|
21
|
-
"description": "Image
|
21
|
+
"description": "Image to analyze",
|
22
|
+
"schema": {
|
23
|
+
"type": "string"
|
24
|
+
}
|
25
|
+
},
|
26
|
+
{
|
27
|
+
"name": "base",
|
28
|
+
"summary": "Base",
|
29
|
+
"required": false,
|
30
|
+
"description": "Base model (mobilenet_v1, mobilenet_v2, or lite_mobilenet_v2)",
|
22
31
|
"schema": {
|
23
32
|
"type": "string"
|
24
33
|
}
|
@@ -30,14 +39,23 @@
|
|
30
39
|
"/faceDetection": {
|
31
40
|
"post": {
|
32
41
|
"tags": ["service"],
|
33
|
-
"summary": "
|
34
|
-
"description": "
|
42
|
+
"summary": "Detects faces",
|
43
|
+
"description": "Detects faces in an image",
|
35
44
|
"parameters": [
|
36
45
|
{
|
37
46
|
"name": "image",
|
38
47
|
"summary": "Image",
|
39
48
|
"required": true,
|
40
|
-
"description": "Image
|
49
|
+
"description": "Image to analyze",
|
50
|
+
"schema": {
|
51
|
+
"type": "string"
|
52
|
+
}
|
53
|
+
},
|
54
|
+
{
|
55
|
+
"name": "runtime",
|
56
|
+
"summary": "Runtime environment",
|
57
|
+
"required": false,
|
58
|
+
"description": "Runtime environment (mediapipe or tfjs)",
|
41
59
|
"schema": {
|
42
60
|
"type": "string"
|
43
61
|
}
|