@lota-sdk/shared 0.1.17 → 0.1.19
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json
CHANGED
package/src/constants/model.ts
CHANGED
|
@@ -4,31 +4,31 @@ export const OPENROUTER_TEAM_AGENT_MODEL_ID = 'openrouter/google/gemini-3.1-pro-
|
|
|
4
4
|
export const OPENROUTER_STRUCTURED_HELPER_MODEL_ID = 'openrouter/google/gemini-3-flash-preview' as const
|
|
5
5
|
export const OPENROUTER_DELEGATED_REASONING_MODEL_ID = 'openrouter/google/gemini-3-flash-preview' as const
|
|
6
6
|
export const OPENROUTER_WEB_RESEARCH_MODEL_ID = 'openrouter/stepfun/step-3.5-flash' as const
|
|
7
|
-
export const OPENROUTER_FAST_REASONING_MODEL_ID = 'openrouter/
|
|
8
|
-
export const OPENROUTER_STRUCTURED_REASONING_MODEL_ID = 'openrouter/
|
|
7
|
+
export const OPENROUTER_FAST_REASONING_MODEL_ID = 'openrouter/qwen/qwen3.5-flash-02-23' as const
|
|
8
|
+
export const OPENROUTER_STRUCTURED_REASONING_MODEL_ID = 'openrouter/google/gemini-3.1-pro-preview' as const
|
|
9
9
|
|
|
10
|
-
export const
|
|
10
|
+
export const AI_GATEWAY_REASONING_SUMMARY_LEVEL = 'detailed' as const
|
|
11
11
|
|
|
12
12
|
export const OPENAI_HIGH_REASONING_PROVIDER_OPTIONS = {
|
|
13
|
-
openai: { forceReasoning: true, reasoningEffort: 'high', reasoningSummary:
|
|
13
|
+
openai: { forceReasoning: true, reasoningEffort: 'high', reasoningSummary: AI_GATEWAY_REASONING_SUMMARY_LEVEL },
|
|
14
14
|
} as const
|
|
15
15
|
|
|
16
16
|
export const OPENROUTER_HIGH_REASONING_PROVIDER_OPTIONS = {
|
|
17
|
-
openai: { forceReasoning: true, reasoningEffort: 'high', reasoningSummary:
|
|
17
|
+
openai: { forceReasoning: true, reasoningEffort: 'high', reasoningSummary: AI_GATEWAY_REASONING_SUMMARY_LEVEL },
|
|
18
18
|
} as const
|
|
19
19
|
|
|
20
20
|
export const OPENROUTER_XHIGH_REASONING_PROVIDER_OPTIONS = {
|
|
21
|
-
openai: { forceReasoning: true, reasoningEffort: 'xhigh', reasoningSummary:
|
|
21
|
+
openai: { forceReasoning: true, reasoningEffort: 'xhigh', reasoningSummary: AI_GATEWAY_REASONING_SUMMARY_LEVEL },
|
|
22
22
|
} as const
|
|
23
23
|
|
|
24
24
|
export const OPENROUTER_MEDIUM_REASONING_PROVIDER_OPTIONS = {
|
|
25
|
-
openai: { forceReasoning: true, reasoningEffort: 'medium', reasoningSummary:
|
|
25
|
+
openai: { forceReasoning: true, reasoningEffort: 'medium', reasoningSummary: AI_GATEWAY_REASONING_SUMMARY_LEVEL },
|
|
26
26
|
} as const
|
|
27
27
|
|
|
28
28
|
export const OPENROUTER_LOW_REASONING_PROVIDER_OPTIONS = {
|
|
29
|
-
openai: { forceReasoning: true, reasoningEffort: 'low', reasoningSummary:
|
|
29
|
+
openai: { forceReasoning: true, reasoningEffort: 'low', reasoningSummary: AI_GATEWAY_REASONING_SUMMARY_LEVEL },
|
|
30
30
|
} as const
|
|
31
31
|
|
|
32
32
|
export const OPENROUTER_MINIMAL_REASONING_PROVIDER_OPTIONS = {
|
|
33
|
-
openai: { forceReasoning: true, reasoningEffort: 'minimal', reasoningSummary:
|
|
33
|
+
openai: { forceReasoning: true, reasoningEffort: 'minimal', reasoningSummary: AI_GATEWAY_REASONING_SUMMARY_LEVEL },
|
|
34
34
|
} as const
|
|
@@ -1,4 +1,11 @@
|
|
|
1
|
-
import type {
|
|
1
|
+
import type {
|
|
2
|
+
LanguageModel,
|
|
3
|
+
Output,
|
|
4
|
+
PrepareStepFunction,
|
|
5
|
+
StopCondition,
|
|
6
|
+
ToolLoopAgentOnFinishCallback,
|
|
7
|
+
ToolSet,
|
|
8
|
+
} from 'ai'
|
|
2
9
|
|
|
3
10
|
export type ChatMode = 'direct' | 'workstreamMode' | 'fixedWorkstreamMode'
|
|
4
11
|
|
|
@@ -11,7 +18,7 @@ export interface CreateRoutedAgentOptions<TTools extends ToolSet = ToolSet> {
|
|
|
11
18
|
prepareStep?: PrepareStepFunction<TTools>
|
|
12
19
|
maxRetries?: number
|
|
13
20
|
output?: Output.Output
|
|
14
|
-
modelOverride?: { model:
|
|
21
|
+
modelOverride?: { model: LanguageModel; providerOptions?: Record<string, unknown> }
|
|
15
22
|
onFinish?: ToolLoopAgentOnFinishCallback<TTools>
|
|
16
23
|
}
|
|
17
24
|
|