@digipair/skill-ollama 0.95.7 → 0.97.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@digipair/skill-ollama",
3
- "version": "0.95.7",
3
+ "version": "0.97.0",
4
4
  "keywords": [
5
5
  "digipair",
6
6
  "service",
package/schema.fr.json CHANGED
@@ -51,7 +51,47 @@
51
51
  }
52
52
  }
53
53
  ],
54
- "x-events": []
54
+ "x-events": [],
55
+ "responses": {
56
+ "200": {
57
+ "description": "Instance de modèle LLM Ollama créée avec succès",
58
+ "content": {
59
+ "application/json": {
60
+ "schema": {
61
+ "type": "object",
62
+ "description": "Instance de modèle LangChain ChatOllama prête pour la conversation",
63
+ "properties": {
64
+ "model": {
65
+ "type": "string",
66
+ "description": "Nom du modèle (ex: mistral, llama2, codellama)"
67
+ },
68
+ "temperature": {
69
+ "type": "number",
70
+ "description": "Paramètre de température du modèle"
71
+ },
72
+ "baseUrl": {
73
+ "type": "string",
74
+ "description": "URL du serveur Ollama"
75
+ },
76
+ "keepAlive": {
77
+ "type": "string",
78
+ "description": "Durée de maintien du modèle actif"
79
+ }
80
+ }
81
+ }
82
+ }
83
+ }
84
+ },
85
+ "400": {
86
+ "description": "Requête invalide - Nom de modèle ou paramètres invalides"
87
+ },
88
+ "503": {
89
+ "description": "Service indisponible - Serveur Ollama non accessible"
90
+ },
91
+ "500": {
92
+ "description": "Erreur serveur - Erreur de chargement ou d'initialisation du modèle"
93
+ }
94
+ }
55
95
  }
56
96
  },
57
97
  "/embeddings": {
@@ -88,7 +128,43 @@
88
128
  }
89
129
  }
90
130
  ],
91
- "x-events": []
131
+ "x-events": [],
132
+ "responses": {
133
+ "200": {
134
+ "description": "Instance de modèle d'embeddings Ollama créée avec succès",
135
+ "content": {
136
+ "application/json": {
137
+ "schema": {
138
+ "type": "object",
139
+ "description": "Instance de modèle LangChain OllamaEmbeddings prête pour l'embedding de texte",
140
+ "properties": {
141
+ "model": {
142
+ "type": "string",
143
+ "description": "Nom du modèle d'embeddings (ex: nomic-embed-text, all-minilm)"
144
+ },
145
+ "baseUrl": {
146
+ "type": "string",
147
+ "description": "URL du serveur Ollama"
148
+ },
149
+ "dimensions": {
150
+ "type": "number",
151
+ "description": "Dimensions du vecteur d'embedding"
152
+ }
153
+ }
154
+ }
155
+ }
156
+ }
157
+ },
158
+ "400": {
159
+ "description": "Requête invalide - Nom de modèle d'embeddings ou paramètres invalides"
160
+ },
161
+ "503": {
162
+ "description": "Service indisponible - Serveur Ollama non accessible"
163
+ },
164
+ "500": {
165
+ "description": "Erreur serveur - Erreur de chargement ou d'initialisation du modèle d'embeddings"
166
+ }
167
+ }
92
168
  }
93
169
  }
94
170
  },
package/schema.json CHANGED
@@ -51,7 +51,47 @@
51
51
  }
52
52
  }
53
53
  ],
54
- "x-events": []
54
+ "x-events": [],
55
+ "responses": {
56
+ "200": {
57
+ "description": "Ollama LLM model instance created successfully",
58
+ "content": {
59
+ "application/json": {
60
+ "schema": {
61
+ "type": "object",
62
+ "description": "LangChain ChatOllama model instance ready for conversation",
63
+ "properties": {
64
+ "model": {
65
+ "type": "string",
66
+ "description": "Model name (e.g., mistral, llama2, codellama)"
67
+ },
68
+ "temperature": {
69
+ "type": "number",
70
+ "description": "Model temperature setting"
71
+ },
72
+ "baseUrl": {
73
+ "type": "string",
74
+ "description": "Ollama server URL"
75
+ },
76
+ "keepAlive": {
77
+ "type": "string",
78
+ "description": "Model keep-alive duration"
79
+ }
80
+ }
81
+ }
82
+ }
83
+ }
84
+ },
85
+ "400": {
86
+ "description": "Bad request - Invalid model name or parameters"
87
+ },
88
+ "503": {
89
+ "description": "Service unavailable - Ollama server not reachable"
90
+ },
91
+ "500": {
92
+ "description": "Server error - Model loading or initialization error"
93
+ }
94
+ }
55
95
  }
56
96
  },
57
97
  "/embeddings": {
@@ -88,7 +128,43 @@
88
128
  }
89
129
  }
90
130
  ],
91
- "x-events": []
131
+ "x-events": [],
132
+ "responses": {
133
+ "200": {
134
+ "description": "Ollama embeddings model instance created successfully",
135
+ "content": {
136
+ "application/json": {
137
+ "schema": {
138
+ "type": "object",
139
+ "description": "LangChain OllamaEmbeddings model instance ready for text embedding",
140
+ "properties": {
141
+ "model": {
142
+ "type": "string",
143
+ "description": "Embeddings model name (e.g., nomic-embed-text, all-minilm)"
144
+ },
145
+ "baseUrl": {
146
+ "type": "string",
147
+ "description": "Ollama server URL"
148
+ },
149
+ "dimensions": {
150
+ "type": "number",
151
+ "description": "Embedding vector dimensions"
152
+ }
153
+ }
154
+ }
155
+ }
156
+ }
157
+ },
158
+ "400": {
159
+ "description": "Bad request - Invalid embeddings model name or parameters"
160
+ },
161
+ "503": {
162
+ "description": "Service unavailable - Ollama server not reachable"
163
+ },
164
+ "500": {
165
+ "description": "Server error - Embeddings model loading or initialization error"
166
+ }
167
+ }
92
168
  }
93
169
  }
94
170
  },