@runnerpro/backend 1.15.7 → 1.16.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -325,7 +325,7 @@ const sendEmoji = (req, res, { sendNotification, firebaseMessaging, isClient })
325
325
  sendNotification({
326
326
  firebaseMessaging,
327
327
  idCliente,
328
- body: `Rubén ha reaccionado a un mensaje '${emoji}'`,
328
+ body: `El entrenador ha reaccionado a un mensaje '${emoji}'`,
329
329
  screen: common_1.NOTIFICATION_SCREEN_TYPES.CHAT,
330
330
  });
331
331
  }
@@ -522,7 +522,7 @@ Sigue estas reglas estrictamente:
522
522
  **TAREA ACTUAL**
523
523
  **Texto Original:** ${text}
524
524
  **Texto Corregido:**
525
- `);
525
+ `, 'LITE');
526
526
  textCorregido = textCorregido.replace(/[¡¿]/g, '');
527
527
  // Eliminar el último punto de las frases
528
528
  textCorregido = textCorregido.trim().replace(/\.$/, '');
@@ -11,7 +11,7 @@ var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, ge
11
11
  Object.defineProperty(exports, "__esModule", { value: true });
12
12
  exports.sendPrompt = void 0;
13
13
  const vertexai_1 = require("@google-cloud/vertexai");
14
- function sendPrompt(prompt) {
14
+ function sendPrompt(prompt, model = 'FLASH') {
15
15
  var _a, _b, _c;
16
16
  return __awaiter(this, void 0, void 0, function* () {
17
17
  // Inicializa el cliente de Vertex AI
@@ -20,10 +20,16 @@ function sendPrompt(prompt) {
20
20
  location: 'europe-west8', // Reemplaza con tu región
21
21
  });
22
22
  // Selecciona el modelo de Gemini más rápido
23
- const model = 'gemini-2.0-flash-lite';
23
+ let modelGemini;
24
+ if (model === 'LITE')
25
+ modelGemini = 'gemini-2.5-flash-lite';
26
+ else if (model === 'FLASH')
27
+ modelGemini = 'gemini-2.5-flash';
28
+ else if (model === 'PRO')
29
+ modelGemini = 'gemini-2.5-pro';
24
30
  // Configura el modelo generativo con parámetros optimizados para velocidad
25
31
  const generativeModel = vertex_ai.preview.getGenerativeModel({
26
- model: model,
32
+ model: modelGemini,
27
33
  generationConfig: {
28
34
  temperature: 0.3,
29
35
  topP: 0.8,
@@ -1,3 +1,3 @@
1
- declare function sendPrompt(prompt: string): Promise<string>;
1
+ declare function sendPrompt(prompt: string, model?: string): Promise<string>;
2
2
  export { sendPrompt };
3
3
  //# sourceMappingURL=index.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../../src/prompt/index.ts"],"names":[],"mappings":"AAEA,iBAAe,UAAU,CAAC,MAAM,EAAE,MAAM,mBAgCvC;AAED,OAAO,EAAE,UAAU,EAAE,CAAC"}
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../../../src/prompt/index.ts"],"names":[],"mappings":"AAEA,iBAAe,UAAU,CAAC,MAAM,EAAE,MAAM,EAAE,KAAK,GAAE,MAAgB,mBAsChE;AAED,OAAO,EAAE,UAAU,EAAE,CAAC"}
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@runnerpro/backend",
3
- "version": "1.15.7",
3
+ "version": "1.16.0",
4
4
  "description": "A collection of common backend functions",
5
5
  "exports": {
6
6
  ".": "./lib/cjs/index.js"