@agi-cli/server 0.1.139 → 0.1.140
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +3 -3
- package/src/openapi/schemas.ts +3 -3
- package/src/routes/config/models.ts +3 -3
- package/src/routes/git/commit.ts +1 -1
- package/src/routes/git/schemas.ts +1 -1
- package/src/routes/session-messages.ts +2 -2
- package/src/runtime/agent/runner-setup.ts +14 -29
- package/src/runtime/message/compaction-limits.ts +5 -2
- package/src/runtime/message/history-builder.ts +1 -1
- package/src/runtime/message/service.ts +3 -3
- package/src/runtime/session/branch.ts +2 -2
- package/src/runtime/session/db-operations.ts +17 -25
- package/src/runtime/session/queue.ts +1 -1
- package/src/runtime/stream/finish-handler.ts +3 -3
- package/src/runtime/stream/step-finish.ts +2 -2
- package/src/tools/database/get-parent-session.ts +1 -1
- package/src/tools/database/get-session-context.ts +1 -1
- package/src/tools/database/present-session-links.ts +1 -1
- package/src/tools/database/query-messages.ts +1 -1
- package/src/tools/database/query-sessions.ts +1 -1
- package/src/tools/database/search-history.ts +1 -1
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@agi-cli/server",
|
|
3
|
-
"version": "0.1.
|
|
3
|
+
"version": "0.1.140",
|
|
4
4
|
"description": "HTTP API server for AGI CLI",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "./src/index.ts",
|
|
@@ -29,8 +29,8 @@
|
|
|
29
29
|
"typecheck": "tsc --noEmit"
|
|
30
30
|
},
|
|
31
31
|
"dependencies": {
|
|
32
|
-
"@agi-cli/sdk": "0.1.
|
|
33
|
-
"@agi-cli/database": "0.1.
|
|
32
|
+
"@agi-cli/sdk": "0.1.140",
|
|
33
|
+
"@agi-cli/database": "0.1.140",
|
|
34
34
|
"drizzle-orm": "^0.44.5",
|
|
35
35
|
"hono": "^4.9.9",
|
|
36
36
|
"zod": "^4.1.8"
|
package/src/openapi/schemas.ts
CHANGED
|
@@ -91,8 +91,8 @@ export const schemas = {
|
|
|
91
91
|
createdAt: { type: 'integer', format: 'int64' },
|
|
92
92
|
completedAt: { type: 'integer', format: 'int64', nullable: true },
|
|
93
93
|
latencyMs: { type: 'integer', nullable: true },
|
|
94
|
-
|
|
95
|
-
|
|
94
|
+
inputTokens: { type: 'integer', nullable: true },
|
|
95
|
+
outputTokens: { type: 'integer', nullable: true },
|
|
96
96
|
totalTokens: { type: 'integer', nullable: true },
|
|
97
97
|
error: { type: 'string', nullable: true },
|
|
98
98
|
},
|
|
@@ -206,7 +206,7 @@ export const schemas = {
|
|
|
206
206
|
id: { type: 'string' },
|
|
207
207
|
label: { type: 'string' },
|
|
208
208
|
toolCall: { type: 'boolean' },
|
|
209
|
-
|
|
209
|
+
reasoningText: { type: 'boolean' },
|
|
210
210
|
},
|
|
211
211
|
required: ['id', 'label'],
|
|
212
212
|
},
|
|
@@ -59,7 +59,7 @@ export function registerModelsRoutes(app: Hono) {
|
|
|
59
59
|
id: m.id,
|
|
60
60
|
label: m.label || m.id,
|
|
61
61
|
toolCall: m.toolCall,
|
|
62
|
-
|
|
62
|
+
reasoningText: m.reasoningText,
|
|
63
63
|
vision: m.modalities?.input?.includes('image') ?? false,
|
|
64
64
|
})),
|
|
65
65
|
default: getDefault(
|
|
@@ -97,7 +97,7 @@ export function registerModelsRoutes(app: Hono) {
|
|
|
97
97
|
id: string;
|
|
98
98
|
label: string;
|
|
99
99
|
toolCall?: boolean;
|
|
100
|
-
|
|
100
|
+
reasoningText?: boolean;
|
|
101
101
|
}>;
|
|
102
102
|
}
|
|
103
103
|
> = {};
|
|
@@ -122,7 +122,7 @@ export function registerModelsRoutes(app: Hono) {
|
|
|
122
122
|
id: m.id,
|
|
123
123
|
label: m.label || m.id,
|
|
124
124
|
toolCall: m.toolCall,
|
|
125
|
-
|
|
125
|
+
reasoningText: m.reasoningText,
|
|
126
126
|
vision: m.modalities?.input?.includes('image') ?? false,
|
|
127
127
|
})),
|
|
128
128
|
};
|
package/src/routes/git/commit.ts
CHANGED
|
@@ -122,7 +122,7 @@ export function registerSessionMessagesRoutes(app: Hono) {
|
|
|
122
122
|
typeOf: typeof userContext,
|
|
123
123
|
});
|
|
124
124
|
|
|
125
|
-
const reasoning = body?.
|
|
125
|
+
const reasoning = body?.reasoningText === true;
|
|
126
126
|
|
|
127
127
|
// Validate model capabilities if tools are allowed for this agent
|
|
128
128
|
const wantsToolCalls = true; // agent toolset may be non-empty
|
|
@@ -156,7 +156,7 @@ export function registerSessionMessagesRoutes(app: Hono) {
|
|
|
156
156
|
content,
|
|
157
157
|
oneShot: Boolean(body?.oneShot),
|
|
158
158
|
userContext,
|
|
159
|
-
reasoning,
|
|
159
|
+
reasoningText: reasoning,
|
|
160
160
|
images,
|
|
161
161
|
files,
|
|
162
162
|
});
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import { loadConfig,
|
|
1
|
+
import { loadConfig, getUnderlyingProviderKey } from '@agi-cli/sdk';
|
|
2
2
|
import { getDb } from '@agi-cli/database';
|
|
3
3
|
import { sessions } from '@agi-cli/database/schema';
|
|
4
4
|
import { eq } from 'drizzle-orm';
|
|
@@ -40,16 +40,6 @@ export interface SetupResult {
|
|
|
40
40
|
|
|
41
41
|
const THINKING_BUDGET = 16000;
|
|
42
42
|
|
|
43
|
-
function getSolforgeUnderlyingProvider(
|
|
44
|
-
model: string,
|
|
45
|
-
): 'anthropic' | 'openai' | null {
|
|
46
|
-
const entry = catalog.solforge?.models?.find((m) => m.id === model);
|
|
47
|
-
const npm = entry?.provider?.npm;
|
|
48
|
-
if (npm === '@ai-sdk/anthropic') return 'anthropic';
|
|
49
|
-
if (npm === '@ai-sdk/openai') return 'openai';
|
|
50
|
-
return null;
|
|
51
|
-
}
|
|
52
|
-
|
|
53
43
|
export async function setupRunner(opts: RunOpts): Promise<SetupResult> {
|
|
54
44
|
const cfgTimer = time('runner:loadConfig+db');
|
|
55
45
|
const cfg = await loadConfig(opts.projectRoot);
|
|
@@ -231,36 +221,31 @@ export async function setupRunner(opts: RunOpts): Promise<SetupResult> {
|
|
|
231
221
|
const providerOptions: Record<string, unknown> = {};
|
|
232
222
|
let effectiveMaxOutputTokens = maxOutputTokens;
|
|
233
223
|
|
|
234
|
-
if (opts.
|
|
235
|
-
|
|
224
|
+
if (opts.reasoningText) {
|
|
225
|
+
const underlyingProvider = getUnderlyingProviderKey(
|
|
226
|
+
opts.provider,
|
|
227
|
+
opts.model,
|
|
228
|
+
);
|
|
229
|
+
|
|
230
|
+
if (underlyingProvider === 'anthropic') {
|
|
236
231
|
providerOptions.anthropic = {
|
|
237
232
|
thinking: { type: 'enabled', budgetTokens: THINKING_BUDGET },
|
|
238
233
|
};
|
|
239
234
|
if (maxOutputTokens && maxOutputTokens > THINKING_BUDGET) {
|
|
240
235
|
effectiveMaxOutputTokens = maxOutputTokens - THINKING_BUDGET;
|
|
241
236
|
}
|
|
242
|
-
} else if (
|
|
237
|
+
} else if (underlyingProvider === 'openai') {
|
|
243
238
|
providerOptions.openai = {
|
|
244
239
|
reasoningSummary: 'auto',
|
|
245
240
|
};
|
|
246
|
-
} else if (
|
|
241
|
+
} else if (underlyingProvider === 'google') {
|
|
247
242
|
providerOptions.google = {
|
|
248
243
|
thinkingConfig: { thinkingBudget: THINKING_BUDGET },
|
|
249
244
|
};
|
|
250
|
-
} else if (
|
|
251
|
-
|
|
252
|
-
|
|
253
|
-
|
|
254
|
-
thinking: { type: 'enabled', budgetTokens: THINKING_BUDGET },
|
|
255
|
-
};
|
|
256
|
-
if (maxOutputTokens && maxOutputTokens > THINKING_BUDGET) {
|
|
257
|
-
effectiveMaxOutputTokens = maxOutputTokens - THINKING_BUDGET;
|
|
258
|
-
}
|
|
259
|
-
} else if (underlying === 'openai') {
|
|
260
|
-
providerOptions.openai = {
|
|
261
|
-
reasoningSummary: 'auto',
|
|
262
|
-
};
|
|
263
|
-
}
|
|
245
|
+
} else if (underlyingProvider === 'openai-compatible') {
|
|
246
|
+
providerOptions['openai-compatible'] = {
|
|
247
|
+
reasoningEffort: 'high',
|
|
248
|
+
};
|
|
264
249
|
}
|
|
265
250
|
}
|
|
266
251
|
|
|
@@ -9,7 +9,7 @@ export interface TokenUsage {
|
|
|
9
9
|
output: number;
|
|
10
10
|
cacheRead?: number;
|
|
11
11
|
cacheWrite?: number;
|
|
12
|
-
|
|
12
|
+
reasoningText?: number;
|
|
13
13
|
}
|
|
14
14
|
|
|
15
15
|
export interface ModelLimits {
|
|
@@ -17,7 +17,10 @@ export interface ModelLimits {
|
|
|
17
17
|
output: number;
|
|
18
18
|
}
|
|
19
19
|
|
|
20
|
-
export function isOverflow(
|
|
20
|
+
export function isOverflow(
|
|
21
|
+
tokens: LanguageModelUsage,
|
|
22
|
+
limits: ModelLimits,
|
|
23
|
+
): boolean {
|
|
21
24
|
if (limits.context === 0) return false;
|
|
22
25
|
|
|
23
26
|
const count =
|
|
@@ -23,7 +23,7 @@ type DispatchOptions = {
|
|
|
23
23
|
content: string;
|
|
24
24
|
oneShot?: boolean;
|
|
25
25
|
userContext?: string;
|
|
26
|
-
|
|
26
|
+
reasoningText?: boolean;
|
|
27
27
|
images?: Array<{ data: string; mediaType: string }>;
|
|
28
28
|
files?: Array<{
|
|
29
29
|
type: 'image' | 'pdf' | 'text';
|
|
@@ -47,7 +47,7 @@ export async function dispatchAssistantMessage(
|
|
|
47
47
|
content,
|
|
48
48
|
oneShot,
|
|
49
49
|
userContext,
|
|
50
|
-
|
|
50
|
+
reasoningText,
|
|
51
51
|
images,
|
|
52
52
|
files,
|
|
53
53
|
} = options;
|
|
@@ -181,7 +181,7 @@ export async function dispatchAssistantMessage(
|
|
|
181
181
|
projectRoot: cfg.projectRoot,
|
|
182
182
|
oneShot: Boolean(oneShot),
|
|
183
183
|
userContext,
|
|
184
|
-
|
|
184
|
+
reasoningText,
|
|
185
185
|
isCompactCommand: isCompact,
|
|
186
186
|
compactionContext,
|
|
187
187
|
},
|
|
@@ -115,8 +115,8 @@ export async function createBranch({
|
|
|
115
115
|
createdAt: msg.createdAt,
|
|
116
116
|
completedAt: msg.completedAt,
|
|
117
117
|
latencyMs: msg.latencyMs,
|
|
118
|
-
|
|
119
|
-
|
|
118
|
+
inputTokens: msg.inputTokens,
|
|
119
|
+
outputTokens: msg.outputTokens,
|
|
120
120
|
totalTokens: msg.totalTokens,
|
|
121
121
|
cachedInputTokens: msg.cachedInputTokens,
|
|
122
122
|
cacheCreationInputTokens: msg.cacheCreationInputTokens,
|
|
@@ -27,7 +27,7 @@ export type ProviderMetadata = Record<string, unknown> & {
|
|
|
27
27
|
|
|
28
28
|
export function normalizeUsage(
|
|
29
29
|
usage: UsageData,
|
|
30
|
-
|
|
30
|
+
providerOptions: ProviderMetadata | undefined,
|
|
31
31
|
provider: ProviderId,
|
|
32
32
|
): UsageData {
|
|
33
33
|
const rawInputTokens = Number(usage.inputTokens ?? 0);
|
|
@@ -37,17 +37,17 @@ export function normalizeUsage(
|
|
|
37
37
|
const cachedInputTokens =
|
|
38
38
|
usage.cachedInputTokens != null
|
|
39
39
|
? Number(usage.cachedInputTokens)
|
|
40
|
-
:
|
|
41
|
-
? Number(
|
|
42
|
-
:
|
|
43
|
-
? Number(
|
|
40
|
+
: providerOptions?.openai?.cachedPromptTokens != null
|
|
41
|
+
? Number(providerOptions.openai.cachedPromptTokens)
|
|
42
|
+
: providerOptions?.anthropic?.cacheReadInputTokens != null
|
|
43
|
+
? Number(providerOptions.anthropic.cacheReadInputTokens)
|
|
44
44
|
: undefined;
|
|
45
45
|
|
|
46
46
|
const cacheCreationInputTokens =
|
|
47
47
|
usage.cacheCreationInputTokens != null
|
|
48
48
|
? Number(usage.cacheCreationInputTokens)
|
|
49
|
-
:
|
|
50
|
-
? Number(
|
|
49
|
+
: providerOptions?.anthropic?.cacheCreationInputTokens != null
|
|
50
|
+
? Number(providerOptions.anthropic.cacheCreationInputTokens)
|
|
51
51
|
: undefined;
|
|
52
52
|
|
|
53
53
|
const cachedValue = cachedInputTokens ?? 0;
|
|
@@ -99,18 +99,14 @@ export function resolveUsageProvider(
|
|
|
99
99
|
*/
|
|
100
100
|
export async function updateSessionTokensIncremental(
|
|
101
101
|
usage: UsageData,
|
|
102
|
-
|
|
102
|
+
providerOptions: ProviderMetadata | undefined,
|
|
103
103
|
opts: RunOpts,
|
|
104
104
|
db: Awaited<ReturnType<typeof getDb>>,
|
|
105
105
|
) {
|
|
106
106
|
if (!usage || !db) return;
|
|
107
107
|
|
|
108
108
|
const usageProvider = resolveUsageProvider(opts.provider, opts.model);
|
|
109
|
-
const normalizedUsage = normalizeUsage(
|
|
110
|
-
usage,
|
|
111
|
-
providerMetadata,
|
|
112
|
-
usageProvider,
|
|
113
|
-
);
|
|
109
|
+
const normalizedUsage = normalizeUsage(usage, providerOptions, usageProvider);
|
|
114
110
|
|
|
115
111
|
// Read session totals
|
|
116
112
|
const sessRows = await db
|
|
@@ -134,8 +130,8 @@ export async function updateSessionTokensIncremental(
|
|
|
134
130
|
.where(eq(messages.id, opts.assistantMessageId));
|
|
135
131
|
|
|
136
132
|
const msg = msgRows[0];
|
|
137
|
-
const priorPromptMsg = Number(msg?.
|
|
138
|
-
const priorCompletionMsg = Number(msg?.
|
|
133
|
+
const priorPromptMsg = Number(msg?.inputTokens ?? 0);
|
|
134
|
+
const priorCompletionMsg = Number(msg?.outputTokens ?? 0);
|
|
139
135
|
const priorCachedMsg = Number(msg?.cachedInputTokens ?? 0);
|
|
140
136
|
const priorCacheCreationMsg = Number(msg?.cacheCreationInputTokens ?? 0);
|
|
141
137
|
const priorReasoningMsg = Number(msg?.reasoningTokens ?? 0);
|
|
@@ -231,18 +227,14 @@ export async function updateSessionTokens(
|
|
|
231
227
|
*/
|
|
232
228
|
export async function updateMessageTokensIncremental(
|
|
233
229
|
usage: UsageData,
|
|
234
|
-
|
|
230
|
+
providerOptions: ProviderMetadata | undefined,
|
|
235
231
|
opts: RunOpts,
|
|
236
232
|
db: Awaited<ReturnType<typeof getDb>>,
|
|
237
233
|
) {
|
|
238
234
|
if (!usage || !db) return;
|
|
239
235
|
|
|
240
236
|
const usageProvider = resolveUsageProvider(opts.provider, opts.model);
|
|
241
|
-
const normalizedUsage = normalizeUsage(
|
|
242
|
-
usage,
|
|
243
|
-
providerMetadata,
|
|
244
|
-
usageProvider,
|
|
245
|
-
);
|
|
237
|
+
const normalizedUsage = normalizeUsage(usage, providerOptions, usageProvider);
|
|
246
238
|
|
|
247
239
|
const msgRows = await db
|
|
248
240
|
.select()
|
|
@@ -251,8 +243,8 @@ export async function updateMessageTokensIncremental(
|
|
|
251
243
|
|
|
252
244
|
if (msgRows.length > 0 && msgRows[0]) {
|
|
253
245
|
const msg = msgRows[0];
|
|
254
|
-
const priorPrompt = Number(msg.
|
|
255
|
-
const priorCompletion = Number(msg.
|
|
246
|
+
const priorPrompt = Number(msg.inputTokens ?? 0);
|
|
247
|
+
const priorCompletion = Number(msg.outputTokens ?? 0);
|
|
256
248
|
const priorCached = Number(msg.cachedInputTokens ?? 0);
|
|
257
249
|
const priorCacheCreation = Number(msg.cacheCreationInputTokens ?? 0);
|
|
258
250
|
const priorReasoning = Number(msg.reasoningTokens ?? 0);
|
|
@@ -287,8 +279,8 @@ export async function updateMessageTokensIncremental(
|
|
|
287
279
|
await db
|
|
288
280
|
.update(messages)
|
|
289
281
|
.set({
|
|
290
|
-
|
|
291
|
-
|
|
282
|
+
inputTokens: cumPrompt,
|
|
283
|
+
outputTokens: cumCompletion,
|
|
292
284
|
totalTokens: cumTotal,
|
|
293
285
|
cachedInputTokens: cumCached,
|
|
294
286
|
cacheCreationInputTokens: cumCacheCreation,
|
|
@@ -73,8 +73,8 @@ export function createFinishHandler(
|
|
|
73
73
|
|
|
74
74
|
const usage = sessRows[0]
|
|
75
75
|
? {
|
|
76
|
-
inputTokens: Number(sessRows[0].
|
|
77
|
-
outputTokens: Number(sessRows[0].
|
|
76
|
+
inputTokens: Number(sessRows[0].inputTokens ?? 0),
|
|
77
|
+
outputTokens: Number(sessRows[0].outputTokens ?? 0),
|
|
78
78
|
totalTokens: Number(sessRows[0].totalTokens ?? 0),
|
|
79
79
|
cachedInputTokens: Number(sessRows[0].cachedInputTokens ?? 0),
|
|
80
80
|
cacheCreationInputTokens: Number(
|
|
@@ -97,7 +97,7 @@ export function createFinishHandler(
|
|
|
97
97
|
try {
|
|
98
98
|
const limits = getModelLimits(opts.provider, opts.model);
|
|
99
99
|
if (limits) {
|
|
100
|
-
const tokenUsage:
|
|
100
|
+
const tokenUsage: LanguageModelUsage = {
|
|
101
101
|
input: usage.inputTokens ?? 0,
|
|
102
102
|
output: usage.outputTokens ?? 0,
|
|
103
103
|
cacheRead:
|
|
@@ -18,13 +18,13 @@ export function createStepFinishHandler(
|
|
|
18
18
|
sharedCtx: ToolAdapterContext,
|
|
19
19
|
updateSessionTokensIncrementalFn: (
|
|
20
20
|
usage: UsageData,
|
|
21
|
-
|
|
21
|
+
providerOptions: ProviderMetadata | undefined,
|
|
22
22
|
opts: RunOpts,
|
|
23
23
|
db: Awaited<ReturnType<typeof getDb>>,
|
|
24
24
|
) => Promise<void>,
|
|
25
25
|
updateMessageTokensIncrementalFn: (
|
|
26
26
|
usage: UsageData,
|
|
27
|
-
|
|
27
|
+
providerOptions: ProviderMetadata | undefined,
|
|
28
28
|
opts: RunOpts,
|
|
29
29
|
db: Awaited<ReturnType<typeof getDb>>,
|
|
30
30
|
) => Promise<void>,
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import { tool } from 'ai';
|
|
2
|
-
import { z } from 'zod';
|
|
2
|
+
import { z } from 'zod/v3';
|
|
3
3
|
import { getDb } from '@agi-cli/database';
|
|
4
4
|
import { sessions, messages, messageParts } from '@agi-cli/database/schema';
|
|
5
5
|
import { eq, desc, asc, gte, lte, and, like, count, sql } from 'drizzle-orm';
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import { tool } from 'ai';
|
|
2
|
-
import { z } from 'zod';
|
|
2
|
+
import { z } from 'zod/v3';
|
|
3
3
|
import { getDb } from '@agi-cli/database';
|
|
4
4
|
import { sessions, messages, messageParts } from '@agi-cli/database/schema';
|
|
5
5
|
import { eq, desc, asc, like, and, sql } from 'drizzle-orm';
|