@lota-sdk/core 0.1.44 → 0.1.45

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@lota-sdk/core",
3
- "version": "0.1.44",
3
+ "version": "0.1.45",
4
4
  "type": "module",
5
5
  "main": "./src/index.ts",
6
6
  "types": "./src/index.ts",
@@ -32,7 +32,7 @@
32
32
  "@chat-adapter/slack": "^4.23.0",
33
33
  "@chat-adapter/state-ioredis": "^4.23.0",
34
34
  "@logtape/logtape": "^2.0.5",
35
- "@lota-sdk/shared": "0.1.44",
35
+ "@lota-sdk/shared": "0.1.45",
36
36
  "@mendable/firecrawl-js": "^4.18.0",
37
37
  "@surrealdb/node": "^3.0.3",
38
38
  "ai": "^6.0.141",
@@ -4,7 +4,6 @@ import { z } from 'zod'
4
4
  import { aiGatewayChatModel } from '../ai-gateway/ai-gateway'
5
5
  import { buildAiGatewayDirectCacheHeaders } from '../ai-gateway/cache-headers'
6
6
  import { agentDescriptions, agentDisplayNames, routerModelId } from '../config/agent-defaults'
7
- import { OPENROUTER_FAST_REASONING_MODEL_ID } from '../config/model-constants'
8
7
 
9
8
  // ---------------------------------------------------------------------------
10
9
  // Schemas
@@ -96,16 +95,12 @@ Format: {"done":true} or {"done":false,"agentId":"<id>","routingContext":"<1-sen
96
95
  // ---------------------------------------------------------------------------
97
96
 
98
97
  function createRouterAgent(systemPrompt: string) {
99
- const modelId = routerModelId ?? OPENROUTER_FAST_REASONING_MODEL_ID
100
- // Router needs plain JSON output, not reasoning tokens
101
- const providerOptions = routerModelId
102
- ? { openai: { provider: { order: ['groq'], allow_fallbacks: true } } }
103
- : undefined
98
+ const modelId = routerModelId ?? 'openai/gpt-5.4-nano'
104
99
  return new ToolLoopAgent({
105
100
  id: 'workstream-router',
106
101
  model: aiGatewayChatModel(modelId),
107
102
  headers: buildAiGatewayDirectCacheHeaders('workstream-router'),
108
- providerOptions,
103
+ providerOptions: { openai: { reasoningEffort: 'high' } },
109
104
  instructions: systemPrompt,
110
105
  maxOutputTokens: 256,
111
106
  })