@okf/ootils 1.6.9 → 1.6.11

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/node.mjs CHANGED
@@ -40,6 +40,256 @@ var __toESM = (mod, isNodeMode, target) => (target = mod != null ? __create(__ge
40
40
  var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
41
41
  var __publicField = (obj, key, value) => __defNormalProp(obj, typeof key !== "symbol" ? key + "" : key, value);
42
42
 
43
+ // src/bullmq/GLOBAL_BULLMQ_CONFIG.js
44
+ var GLOBAL_BULLMQ_CONFIG_exports = {};
45
+ __export(GLOBAL_BULLMQ_CONFIG_exports, {
46
+ BASE_BULLMQ_CONFIG: () => BASE_BULLMQ_CONFIG
47
+ });
48
+ var BASE_BULLMQ_CONFIG;
49
+ var init_GLOBAL_BULLMQ_CONFIG = __esm({
50
+ "src/bullmq/GLOBAL_BULLMQ_CONFIG.js"() {
51
+ "use strict";
52
+ BASE_BULLMQ_CONFIG = {
53
+ PLUGIN__MAD_USERS_SYNC_QUEUE: {
54
+ id: "plugin--mad-users-sync-queue",
55
+ queueConfig: {
56
+ defaultJobOptions: {
57
+ backoff: {
58
+ type: "exponential",
59
+ delay: 2e3
60
+ },
61
+ attempts: 3,
62
+ removeOnComplete: 30,
63
+ removeOnFail: 100
64
+ },
65
+ streams: {
66
+ events: {
67
+ maxLen: 10
68
+ // Keep very low, cuz we dont really use historical stream events as of now
69
+ }
70
+ }
71
+ },
72
+ workerConfig: {
73
+ concurrency: 1,
74
+ // Process jobs one at a time to avoid race conditions
75
+ limiter: {
76
+ max: 5,
77
+ // Max 10 jobs per...
78
+ duration: 6e4
79
+ // ...60 seconds (rate limiting)
80
+ }
81
+ }
82
+ },
83
+ // Chunk Processing Queue
84
+ CREATE_CHUNKS_QUEUE: {
85
+ id: "create-chunks-queue",
86
+ queueConfig: {
87
+ defaultJobOptions: {
88
+ backoff: {
89
+ type: "exponential",
90
+ delay: 2e3
91
+ },
92
+ attempts: 3,
93
+ removeOnComplete: 30,
94
+ removeOnFail: 100
95
+ },
96
+ streams: {
97
+ events: {
98
+ maxLen: 10
99
+ // Keep very low, cuz we dont really use historical stream events as of now
100
+ }
101
+ }
102
+ },
103
+ workerConfig: {
104
+ concurrency: 10,
105
+ // Process 10 jobs at once for chunk processing
106
+ limiter: {
107
+ max: 100,
108
+ // Max 5 jobs per...
109
+ duration: 6e4
110
+ // ...60 seconds (higher throughput for chunking)
111
+ }
112
+ }
113
+ },
114
+ CREATE_ANNOS_QUEUE: {
115
+ id: "create-annos-queue",
116
+ queueConfig: {
117
+ defaultJobOptions: {
118
+ backoff: {
119
+ type: "exponential",
120
+ delay: 2e3
121
+ },
122
+ attempts: 3,
123
+ removeOnComplete: 30,
124
+ removeOnFail: 100
125
+ },
126
+ streams: {
127
+ events: {
128
+ maxLen: 10
129
+ // Keep very low, cuz we dont really use historical stream events as of now
130
+ }
131
+ }
132
+ },
133
+ workerConfig: {
134
+ concurrency: 10,
135
+ // Process 10 jobs at once for chunk processing
136
+ limiter: {
137
+ max: 100,
138
+ // Max 50 jobs per...
139
+ duration: 6e4
140
+ // ...60 seconds (higher throughput for chunking)
141
+ }
142
+ }
143
+ },
144
+ CONTENT_ENHANCE_AND_EMBED_QUEUE: {
145
+ id: "content-enhance-and-embed-queue",
146
+ queueConfig: {
147
+ defaultJobOptions: {
148
+ backoff: {
149
+ type: "exponential",
150
+ delay: 2e3
151
+ },
152
+ attempts: 3,
153
+ removeOnComplete: 30,
154
+ removeOnFail: 100
155
+ },
156
+ streams: {
157
+ events: {
158
+ maxLen: 10
159
+ // Keep very low, cuz we dont really use historical stream events as of now
160
+ }
161
+ }
162
+ },
163
+ workerConfig: {
164
+ concurrency: 1,
165
+ limiter: {
166
+ max: 100,
167
+ // Max 50 jobs per...
168
+ duration: 6e4
169
+ // ...60 seconds (higher throughput for chunking)
170
+ }
171
+ }
172
+ },
173
+ DIRECT_DATA_IMPORT_QUEUE: {
174
+ id: "direct-data-import-queue",
175
+ queueConfig: {
176
+ defaultJobOptions: {
177
+ backoff: {
178
+ type: "exponential",
179
+ delay: 2e3
180
+ },
181
+ attempts: 3,
182
+ removeOnComplete: 30,
183
+ removeOnFail: 100
184
+ },
185
+ streams: {
186
+ events: {
187
+ maxLen: 10
188
+ // Keep very low, cuz we dont really use historical stream events as of now
189
+ }
190
+ }
191
+ },
192
+ workerConfig: {
193
+ concurrency: 1,
194
+ limiter: {
195
+ max: 5,
196
+ // Max 5 jobs per...
197
+ duration: 6e4
198
+ // ...60 seconds (higher throughput for chunking)
199
+ }
200
+ }
201
+ },
202
+ AI_CATEGORIZE_QUEUE: {
203
+ id: "ai-categorize-queue",
204
+ queueConfig: {
205
+ defaultJobOptions: {
206
+ attempts: 3,
207
+ backoff: {
208
+ type: "exponential",
209
+ delay: 2e3
210
+ },
211
+ removeOnComplete: 30,
212
+ removeOnFail: 100
213
+ },
214
+ streams: {
215
+ events: {
216
+ maxLen: 10
217
+ // Keep very low, cuz we dont really use historical stream events as of now
218
+ }
219
+ }
220
+ },
221
+ workerConfig: {
222
+ concurrency: 1,
223
+ limiter: {
224
+ max: 100,
225
+ // (lets always keep this same as content enhance & embed since it comes immediately after)
226
+ duration: 6e4
227
+ // ...60 seconds (higher throughput for chunking)
228
+ }
229
+ }
230
+ },
231
+ ANNOS_ELASTIC_SYNC_QUEUE: {
232
+ id: "annos-elastic-sync-queue",
233
+ queueConfig: {
234
+ defaultJobOptions: {
235
+ attempts: 3,
236
+ backoff: {
237
+ type: "exponential",
238
+ delay: 2e3
239
+ },
240
+ removeOnComplete: 30,
241
+ removeOnFail: 100
242
+ },
243
+ streams: {
244
+ events: {
245
+ maxLen: 10
246
+ // Keep very low, cuz we dont really use historical stream events as of now
247
+ }
248
+ }
249
+ },
250
+ workerConfig: {
251
+ concurrency: 5,
252
+ limiter: {
253
+ max: 100,
254
+ // (lets always keep this same as content enhance & embed since it comes immediately after)
255
+ duration: 6e4
256
+ // ...60 seconds (higher throughput for chunking)
257
+ }
258
+ }
259
+ },
260
+ CHUNKS_ELASTIC_SYNC_QUEUE: {
261
+ id: "chunks-elastic-sync-queue",
262
+ queueConfig: {
263
+ defaultJobOptions: {
264
+ attempts: 3,
265
+ backoff: {
266
+ type: "exponential",
267
+ delay: 2e3
268
+ },
269
+ removeOnComplete: 30,
270
+ removeOnFail: 100
271
+ },
272
+ streams: {
273
+ events: {
274
+ maxLen: 10
275
+ // Keep very low, cuz we dont really use historical stream events as of now
276
+ }
277
+ }
278
+ },
279
+ workerConfig: {
280
+ concurrency: 5,
281
+ limiter: {
282
+ max: 100,
283
+ // (lets always keep this same as content enhance & embed since it comes immediately after)
284
+ duration: 6e4
285
+ // ...60 seconds (higher throughput for chunking)
286
+ }
287
+ }
288
+ }
289
+ };
290
+ }
291
+ });
292
+
43
293
  // src/models/Annotations.ts
44
294
  var Annotations_exports = {};
45
295
  __export(Annotations_exports, {
@@ -1086,256 +1336,6 @@ var require_BaseWorker = __commonJS({
1086
1336
  }
1087
1337
  });
1088
1338
 
1089
- // src/bullmq/GLOBAL_BULLMQ_CONFIG.js
1090
- var GLOBAL_BULLMQ_CONFIG_exports = {};
1091
- __export(GLOBAL_BULLMQ_CONFIG_exports, {
1092
- BASE_BULLMQ_CONFIG: () => BASE_BULLMQ_CONFIG
1093
- });
1094
- var BASE_BULLMQ_CONFIG;
1095
- var init_GLOBAL_BULLMQ_CONFIG = __esm({
1096
- "src/bullmq/GLOBAL_BULLMQ_CONFIG.js"() {
1097
- "use strict";
1098
- BASE_BULLMQ_CONFIG = {
1099
- PLUGIN__MAD_USERS_SYNC_QUEUE: {
1100
- id: "plugin--mad-users-sync-queue",
1101
- queueConfig: {
1102
- defaultJobOptions: {
1103
- backoff: {
1104
- type: "exponential",
1105
- delay: 2e3
1106
- },
1107
- attempts: 3,
1108
- removeOnComplete: 30,
1109
- removeOnFail: 100
1110
- },
1111
- streams: {
1112
- events: {
1113
- maxLen: 10
1114
- // Keep very low, cuz we dont really use historical stream events as of now
1115
- }
1116
- }
1117
- },
1118
- workerConfig: {
1119
- concurrency: 1,
1120
- // Process jobs one at a time to avoid race conditions
1121
- limiter: {
1122
- max: 5,
1123
- // Max 10 jobs per...
1124
- duration: 6e4
1125
- // ...60 seconds (rate limiting)
1126
- }
1127
- }
1128
- },
1129
- // Chunk Processing Queue
1130
- CREATE_CHUNKS_QUEUE: {
1131
- id: "create-chunks-queue",
1132
- queueConfig: {
1133
- defaultJobOptions: {
1134
- backoff: {
1135
- type: "exponential",
1136
- delay: 2e3
1137
- },
1138
- attempts: 3,
1139
- removeOnComplete: 30,
1140
- removeOnFail: 100
1141
- },
1142
- streams: {
1143
- events: {
1144
- maxLen: 10
1145
- // Keep very low, cuz we dont really use historical stream events as of now
1146
- }
1147
- }
1148
- },
1149
- workerConfig: {
1150
- concurrency: 10,
1151
- // Process 10 jobs at once for chunk processing
1152
- limiter: {
1153
- max: 100,
1154
- // Max 5 jobs per...
1155
- duration: 6e4
1156
- // ...60 seconds (higher throughput for chunking)
1157
- }
1158
- }
1159
- },
1160
- CREATE_ANNOS_QUEUE: {
1161
- id: "create-annos-queue",
1162
- queueConfig: {
1163
- defaultJobOptions: {
1164
- backoff: {
1165
- type: "exponential",
1166
- delay: 2e3
1167
- },
1168
- attempts: 3,
1169
- removeOnComplete: 30,
1170
- removeOnFail: 100
1171
- },
1172
- streams: {
1173
- events: {
1174
- maxLen: 10
1175
- // Keep very low, cuz we dont really use historical stream events as of now
1176
- }
1177
- }
1178
- },
1179
- workerConfig: {
1180
- concurrency: 10,
1181
- // Process 10 jobs at once for chunk processing
1182
- limiter: {
1183
- max: 100,
1184
- // Max 50 jobs per...
1185
- duration: 6e4
1186
- // ...60 seconds (higher throughput for chunking)
1187
- }
1188
- }
1189
- },
1190
- CONTENT_ENHANCE_AND_EMBED_QUEUE: {
1191
- id: "content-enhance-and-embed-queue",
1192
- queueConfig: {
1193
- defaultJobOptions: {
1194
- backoff: {
1195
- type: "exponential",
1196
- delay: 2e3
1197
- },
1198
- attempts: 3,
1199
- removeOnComplete: 30,
1200
- removeOnFail: 100
1201
- },
1202
- streams: {
1203
- events: {
1204
- maxLen: 10
1205
- // Keep very low, cuz we dont really use historical stream events as of now
1206
- }
1207
- }
1208
- },
1209
- workerConfig: {
1210
- concurrency: 1,
1211
- limiter: {
1212
- max: 100,
1213
- // Max 50 jobs per...
1214
- duration: 6e4
1215
- // ...60 seconds (higher throughput for chunking)
1216
- }
1217
- }
1218
- },
1219
- DIRECT_DATA_IMPORT_QUEUE: {
1220
- id: "direct-data-import-queue",
1221
- queueConfig: {
1222
- defaultJobOptions: {
1223
- backoff: {
1224
- type: "exponential",
1225
- delay: 2e3
1226
- },
1227
- attempts: 3,
1228
- removeOnComplete: 30,
1229
- removeOnFail: 100
1230
- },
1231
- streams: {
1232
- events: {
1233
- maxLen: 10
1234
- // Keep very low, cuz we dont really use historical stream events as of now
1235
- }
1236
- }
1237
- },
1238
- workerConfig: {
1239
- concurrency: 1,
1240
- limiter: {
1241
- max: 5,
1242
- // Max 5 jobs per...
1243
- duration: 6e4
1244
- // ...60 seconds (higher throughput for chunking)
1245
- }
1246
- }
1247
- },
1248
- AI_CATEGORIZE_QUEUE: {
1249
- id: "ai-categorize-queue",
1250
- queueConfig: {
1251
- defaultJobOptions: {
1252
- attempts: 3,
1253
- backoff: {
1254
- type: "exponential",
1255
- delay: 2e3
1256
- },
1257
- removeOnComplete: 30,
1258
- removeOnFail: 100
1259
- },
1260
- streams: {
1261
- events: {
1262
- maxLen: 10
1263
- // Keep very low, cuz we dont really use historical stream events as of now
1264
- }
1265
- }
1266
- },
1267
- workerConfig: {
1268
- concurrency: 1,
1269
- limiter: {
1270
- max: 100,
1271
- // (lets always keep this same as content enhance & embed since it comes immediately after)
1272
- duration: 6e4
1273
- // ...60 seconds (higher throughput for chunking)
1274
- }
1275
- }
1276
- },
1277
- ANNOS_ELASTIC_SYNC_QUEUE: {
1278
- id: "annos-elastic-sync-queue",
1279
- queueConfig: {
1280
- defaultJobOptions: {
1281
- attempts: 3,
1282
- backoff: {
1283
- type: "exponential",
1284
- delay: 2e3
1285
- },
1286
- removeOnComplete: 30,
1287
- removeOnFail: 100
1288
- },
1289
- streams: {
1290
- events: {
1291
- maxLen: 10
1292
- // Keep very low, cuz we dont really use historical stream events as of now
1293
- }
1294
- }
1295
- },
1296
- workerConfig: {
1297
- concurrency: 5,
1298
- limiter: {
1299
- max: 50,
1300
- // (lets always keep this same as content enhance & embed since it comes immediately after)
1301
- duration: 6e4
1302
- // ...60 seconds (higher throughput for chunking)
1303
- }
1304
- }
1305
- },
1306
- CHUNKS_ELASTIC_SYNC_QUEUE: {
1307
- id: "chunks-elastic-sync-queue",
1308
- queueConfig: {
1309
- defaultJobOptions: {
1310
- attempts: 3,
1311
- backoff: {
1312
- type: "exponential",
1313
- delay: 2e3
1314
- },
1315
- removeOnComplete: 30,
1316
- removeOnFail: 100
1317
- },
1318
- streams: {
1319
- events: {
1320
- maxLen: 10
1321
- // Keep very low, cuz we dont really use historical stream events as of now
1322
- }
1323
- }
1324
- },
1325
- workerConfig: {
1326
- concurrency: 5,
1327
- limiter: {
1328
- max: 50,
1329
- // (lets always keep this same as content enhance & embed since it comes immediately after)
1330
- duration: 6e4
1331
- // ...60 seconds (higher throughput for chunking)
1332
- }
1333
- }
1334
- }
1335
- };
1336
- }
1337
- });
1338
-
1339
1339
  // src/bullmq/GET_GLOBAL_BULLMQ_CONFIG.js
1340
1340
  var require_GET_GLOBAL_BULLMQ_CONFIG = __commonJS({
1341
1341
  "src/bullmq/GET_GLOBAL_BULLMQ_CONFIG.js"(exports, module) {
@@ -1645,6 +1645,9 @@ var _extractBlocksFromSomeBuilders = ({
1645
1645
  }
1646
1646
  };
1647
1647
 
1648
+ // src/universal.ts
1649
+ init_GLOBAL_BULLMQ_CONFIG();
1650
+
1648
1651
  // src/node.ts
1649
1652
  var import_MongoConnector2 = __toESM(require_MongoConnector());
1650
1653
  var import_ElasticSearchConnector = __toESM(require_ElasticSearchConnector());
@@ -1984,6 +1987,7 @@ var export_getTplModelByTenant = import_getModelByTenant2.getTplModelByTenant;
1984
1987
  export {
1985
1988
  AIChat_default as AIChatSchema,
1986
1989
  Annotations_default as AnnotationSchema,
1990
+ BASE_BULLMQ_CONFIG,
1987
1991
  export_BaseProducer as BaseProducer,
1988
1992
  export_BaseWorker as BaseWorker,
1989
1993
  export_ElasticSearchConnector as ElasticSearchConnector,