@digipair/skill-ollama 0.7.3 → 0.8.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/index.cjs.js +38452 -4
- package/index.esm.js +38454 -1
- package/libs/skill-ollama/src/lib/skill-ollama.d.ts +4 -4
- package/package.json +1 -1
- package/schema.json +13 -176
- package/api_chain.cjs.js +0 -480
- package/api_chain.esm.js +0 -478
- package/chat.cjs.js +0 -2015
- package/chat.esm.js +0 -2010
- package/few_shot.cjs.js +0 -888
- package/few_shot.esm.js +0 -885
- package/index.cjs2.js +0 -45897
- package/index.esm2.js +0 -45874
- package/sequential_chain.cjs.js +0 -1092
- package/sequential_chain.esm.js +0 -1089
- package/vector_db_qa.cjs.js +0 -866
- package/vector_db_qa.esm.js +0 -864
@@ -1,5 +1,5 @@
|
|
1
1
|
import { PinsSettings } from '@digipair/engine';
|
2
|
-
import {
|
3
|
-
|
4
|
-
export declare const
|
5
|
-
export declare const
|
2
|
+
import { Ollama } from '@langchain/community/llms/ollama';
|
3
|
+
import { OllamaEmbeddings } from '@langchain/community/embeddings/ollama';
|
4
|
+
export declare const model: (params: any, pinsSettingsList: PinsSettings[], context: any) => Promise<Ollama>;
|
5
|
+
export declare const embeddings: (params: any, pinsSettingsList: PinsSettings[], context: any) => Promise<OllamaEmbeddings>;
|
package/package.json
CHANGED
package/schema.json
CHANGED
@@ -8,25 +8,16 @@
|
|
8
8
|
"x-icon": "🚀"
|
9
9
|
},
|
10
10
|
"paths": {
|
11
|
-
"/
|
11
|
+
"/model": {
|
12
12
|
"post": {
|
13
13
|
"tags": ["service"],
|
14
|
-
"summary": "
|
15
|
-
"description": "
|
14
|
+
"summary": "Modèle LLM Ollama",
|
15
|
+
"description": "Exécution d'un modèle LLM depuis un serveur Ollama",
|
16
16
|
"parameters": [
|
17
17
|
{
|
18
|
-
"name": "
|
19
|
-
"summary": "Prompt",
|
20
|
-
"required": true,
|
21
|
-
"description": "Prompt à exécuter via le model LLM",
|
22
|
-
"schema": {
|
23
|
-
"type": "string"
|
24
|
-
}
|
25
|
-
},
|
26
|
-
{
|
27
|
-
"name": "modelName",
|
18
|
+
"name": "model",
|
28
19
|
"summary": "Nom du model",
|
29
|
-
"required":
|
20
|
+
"required": true,
|
30
21
|
"description": "Nom du model LLM à charger",
|
31
22
|
"schema": {
|
32
23
|
"type": "string"
|
@@ -49,104 +40,31 @@
|
|
49
40
|
"schema": {
|
50
41
|
"type": "string"
|
51
42
|
}
|
52
|
-
},
|
53
|
-
{
|
54
|
-
"name": "schema",
|
55
|
-
"summary": "Schema",
|
56
|
-
"required": false,
|
57
|
-
"description": "Schema JSON des données à extraire",
|
58
|
-
"schema": {
|
59
|
-
"type": "object"
|
60
|
-
}
|
61
43
|
}
|
62
44
|
],
|
63
45
|
"x-events": []
|
64
46
|
}
|
65
47
|
},
|
66
|
-
"/
|
48
|
+
"/embeddings": {
|
67
49
|
"post": {
|
68
50
|
"tags": ["service"],
|
69
|
-
"summary": "
|
70
|
-
"description": "
|
51
|
+
"summary": "Modèle d'embeddings Ollama",
|
52
|
+
"description": "Exécution d'un modèle d'enbeddings depuis un serveur Ollama",
|
71
53
|
"parameters": [
|
72
54
|
{
|
73
|
-
"name": "
|
74
|
-
"summary": "
|
75
|
-
"required": false,
|
76
|
-
"description": "Nom du modèle LLM à utiliser",
|
77
|
-
"schema": {
|
78
|
-
"type": "string"
|
79
|
-
}
|
80
|
-
},
|
81
|
-
{
|
82
|
-
"name": "temperature",
|
83
|
-
"summary": "Temperature",
|
84
|
-
"required": false,
|
85
|
-
"description": "Temperature du model LLM",
|
86
|
-
"schema": {
|
87
|
-
"type": "number"
|
88
|
-
}
|
89
|
-
},
|
90
|
-
{
|
91
|
-
"name": "baseUrl",
|
92
|
-
"summary": "Adresse du serveur",
|
93
|
-
"required": false,
|
94
|
-
"description": "Adresse du serveur Ollama",
|
95
|
-
"schema": {
|
96
|
-
"type": "string"
|
97
|
-
}
|
98
|
-
},
|
99
|
-
{
|
100
|
-
"name": "prompt",
|
101
|
-
"summary": "Prompt",
|
102
|
-
"required": true,
|
103
|
-
"description": "Prompt à exécuter sur le modèle LLM",
|
104
|
-
"schema": {
|
105
|
-
"type": "string"
|
106
|
-
}
|
107
|
-
},
|
108
|
-
{
|
109
|
-
"name": "image",
|
110
|
-
"summary": "Image",
|
55
|
+
"name": "model",
|
56
|
+
"summary": "Nom du model",
|
111
57
|
"required": true,
|
112
|
-
"description": "
|
113
|
-
"schema": {
|
114
|
-
"type": "string"
|
115
|
-
}
|
116
|
-
},
|
117
|
-
{
|
118
|
-
"name": "schema",
|
119
|
-
"summary": "Schema JSON",
|
120
|
-
"required": false,
|
121
|
-
"description": "Schema JSON des données à extraire",
|
122
|
-
"schema": {
|
123
|
-
"type": "object"
|
124
|
-
}
|
125
|
-
}
|
126
|
-
],
|
127
|
-
"x-events": []
|
128
|
-
}
|
129
|
-
},
|
130
|
-
"/summarization": {
|
131
|
-
"post": {
|
132
|
-
"tags": ["service"],
|
133
|
-
"summary": "Résumer un texte",
|
134
|
-
"description": "Résumer un texte via un model LLM",
|
135
|
-
"parameters": [
|
136
|
-
{
|
137
|
-
"name": "modelName",
|
138
|
-
"summary": "Nom du modèle",
|
139
|
-
"required": false,
|
140
|
-
"description": "model name",
|
58
|
+
"description": "Nom du model d'enbeddings à charger",
|
141
59
|
"schema": {
|
142
60
|
"type": "string"
|
143
61
|
}
|
144
62
|
},
|
145
63
|
{
|
146
64
|
"name": "temperature",
|
147
|
-
"summary": "
|
65
|
+
"summary": "Temperature",
|
148
66
|
"required": false,
|
149
|
-
"description": "
|
67
|
+
"description": "Temperature du model d'emdeddings",
|
150
68
|
"schema": {
|
151
69
|
"type": "number"
|
152
70
|
}
|
@@ -159,87 +77,6 @@
|
|
159
77
|
"schema": {
|
160
78
|
"type": "string"
|
161
79
|
}
|
162
|
-
},
|
163
|
-
{
|
164
|
-
"name": "chunkSize",
|
165
|
-
"summary": "Taille des morceaux de texte",
|
166
|
-
"required": false,
|
167
|
-
"description": "Taille des morceaux de texte",
|
168
|
-
"schema": {
|
169
|
-
"type": "number"
|
170
|
-
}
|
171
|
-
},
|
172
|
-
{
|
173
|
-
"name": "type",
|
174
|
-
"summary": "Algorithme",
|
175
|
-
"required": false,
|
176
|
-
"description": "Algorithme à utiliser pour résumer le texte",
|
177
|
-
"schema": {
|
178
|
-
"type": "string"
|
179
|
-
}
|
180
|
-
},
|
181
|
-
{
|
182
|
-
"name": "prompt",
|
183
|
-
"summary": "Texte",
|
184
|
-
"required": true,
|
185
|
-
"description": "Texte à résumer",
|
186
|
-
"schema": {
|
187
|
-
"type": "string"
|
188
|
-
}
|
189
|
-
},
|
190
|
-
{
|
191
|
-
"name": "combineMapPrompt",
|
192
|
-
"summary": "",
|
193
|
-
"required": false,
|
194
|
-
"description": "combineMapPrompt",
|
195
|
-
"schema": {
|
196
|
-
"type": "string"
|
197
|
-
}
|
198
|
-
},
|
199
|
-
{
|
200
|
-
"name": "combinePrompt",
|
201
|
-
"summary": "",
|
202
|
-
"required": false,
|
203
|
-
"description": "combinePrompt",
|
204
|
-
"schema": {
|
205
|
-
"type": "string"
|
206
|
-
}
|
207
|
-
},
|
208
|
-
{
|
209
|
-
"name": "returnIntermediateSteps",
|
210
|
-
"summary": "",
|
211
|
-
"required": false,
|
212
|
-
"description": "returnIntermediateSteps",
|
213
|
-
"schema": {
|
214
|
-
"type": "boolean"
|
215
|
-
}
|
216
|
-
},
|
217
|
-
{
|
218
|
-
"name": "refinePrompt",
|
219
|
-
"summary": "",
|
220
|
-
"required": false,
|
221
|
-
"description": "refinePrompt",
|
222
|
-
"schema": {
|
223
|
-
"type": "string"
|
224
|
-
}
|
225
|
-
},
|
226
|
-
{
|
227
|
-
"name": "questionPrompt",
|
228
|
-
"summary": "",
|
229
|
-
"required": false,
|
230
|
-
"description": "questionPrompt",
|
231
|
-
"schema": {
|
232
|
-
"type": "string"
|
233
|
-
}
|
234
|
-
},
|
235
|
-
{
|
236
|
-
"name": "verbose",
|
237
|
-
"summary": "Debug",
|
238
|
-
"required": false,
|
239
|
-
"description": "Ajoute des logs de debug",
|
240
|
-
"schema": {
|
241
|
-
"type": "boolean"
|
242
|
-
}
|
243
80
|
}
|
244
81
|
],
|
245
82
|
"x-events": []
|