te.js 2.2.1 → 2.2.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "te.js",
3
- "version": "2.2.1",
3
+ "version": "2.2.2",
4
4
  "description": "AI Native Node.js Framework",
5
5
  "type": "module",
6
6
  "main": "te.js",
package/te.js CHANGED
@@ -275,7 +275,7 @@ class Tejas {
275
275
  * @param {string} [config.model] - Model name (e.g. gpt-4o-mini)
276
276
  * @param {'endUser'|'developer'} [config.messageType] - Default message tone
277
277
  * @param {'sync'|'async'} [config.mode] - 'sync' blocks the response until LLM returns (default); 'async' responds immediately with 500 and dispatches LLM result to a channel
278
- * @param {number} [config.timeout] - LLM fetch timeout in milliseconds (default 10000)
278
+ * @param {number} [config.timeout] - LLM fetch timeout in milliseconds (default 20000)
279
279
  * @param {'console'|'log'|'both'} [config.channel] - Output channel for async mode results (default 'console')
280
280
  * @param {string} [config.logFile] - Path to JSONL log file used by 'log' and 'both' channels (default './errors.llm.log')
281
281
  * @param {number} [config.rateLimit] - Max LLM calls per minute across all requests (default 10)
@@ -107,7 +107,7 @@ export function getErrorsLlmConfig() {
107
107
  const timeoutRaw = env('ERRORS_LLM_TIMEOUT') ?? env('LLM_TIMEOUT') ?? '';
108
108
  const timeoutNum = Number(timeoutRaw);
109
109
  const timeout =
110
- !timeoutRaw || isNaN(timeoutNum) || timeoutNum <= 0 ? 10000 : timeoutNum;
110
+ !timeoutRaw || isNaN(timeoutNum) || timeoutNum <= 0 ? 20000 : timeoutNum;
111
111
 
112
112
  const channelRaw = env('ERRORS_LLM_CHANNEL') ?? env('LLM_CHANNEL') ?? '';
113
113