@push.rocks/smartai 2.2.0 → 4.0.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.smartconfig.json +12 -7
- package/dist_ts/00_commitinfo_data.js +1 -1
- package/dist_ts/index.d.ts +5 -1
- package/dist_ts/index.js +3 -1
- package/dist_ts/smartai.auth.openai.d.ts +30 -0
- package/dist_ts/smartai.auth.openai.js +230 -0
- package/dist_ts/smartai.cache.d.ts +31 -0
- package/dist_ts/smartai.cache.js +179 -0
- package/dist_ts/smartai.classes.smartai.js +6 -3
- package/dist_ts/smartai.interfaces.d.ts +49 -3
- package/dist_ts/smartai.middleware.anthropic.d.ts +2 -1
- package/dist_ts/smartai.middleware.anthropic.js +4 -30
- package/dist_ts_openai_chatgpt_auth/index.d.ts +43 -0
- package/dist_ts_openai_chatgpt_auth/index.js +249 -0
- package/package.json +27 -29
- package/readme.hints.md +3 -0
- package/readme.md +86 -3
- package/ts/00_commitinfo_data.ts +1 -1
- package/ts/index.ts +39 -0
- package/ts/smartai.auth.openai.ts +312 -0
- package/ts/smartai.cache.ts +250 -0
- package/ts/smartai.classes.smartai.ts +9 -2
- package/ts/smartai.interfaces.ts +56 -3
- package/ts/smartai.middleware.anthropic.ts +5 -31
- package/ts_openai_chatgpt_auth/index.ts +351 -0
|
@@ -0,0 +1,250 @@
|
|
|
1
|
+
import type { JSONObject, JSONValue, LanguageModelV3Middleware, LanguageModelV3Prompt } from '@ai-sdk/provider';
|
|
2
|
+
import type { TSmartAiProviderOptions } from './smartai.interfaces.js';
|
|
3
|
+
|
|
4
|
+
export type TSmartAiMessageCacheProvider =
|
|
5
|
+
| 'anthropic'
|
|
6
|
+
| 'openrouter'
|
|
7
|
+
| 'bedrock'
|
|
8
|
+
| 'openaiCompatible'
|
|
9
|
+
| 'copilot'
|
|
10
|
+
| 'alibaba';
|
|
11
|
+
|
|
12
|
+
export type TSmartAiCacheRetention = 'ephemeral' | '1h' | 'in_memory' | '24h';
|
|
13
|
+
|
|
14
|
+
export interface ISmartAiCacheOptions {
|
|
15
|
+
/** Provider-specific message cache marker namespace. Usually inferred from the model. */
|
|
16
|
+
provider?: TSmartAiMessageCacheProvider;
|
|
17
|
+
/** Stable session/request key for providers that support request-level prompt cache affinity. */
|
|
18
|
+
key?: string;
|
|
19
|
+
/** Short retention is the default; longer retention is opt-in. */
|
|
20
|
+
retention?: TSmartAiCacheRetention;
|
|
21
|
+
}
|
|
22
|
+
|
|
23
|
+
export type TSmartAiCacheSetting = boolean | 'auto' | ISmartAiCacheOptions;
|
|
24
|
+
|
|
25
|
+
function isObject(input: unknown): input is Record<string, unknown> {
|
|
26
|
+
return typeof input === 'object' && input !== null && !Array.isArray(input);
|
|
27
|
+
}
|
|
28
|
+
|
|
29
|
+
function mergeJsonDefaults(defaults: JSONObject, overrides?: JSONObject): JSONObject {
|
|
30
|
+
const result: JSONObject = { ...defaults };
|
|
31
|
+
|
|
32
|
+
if (!overrides) return result;
|
|
33
|
+
|
|
34
|
+
for (const [key, value] of Object.entries(overrides)) {
|
|
35
|
+
const existing = result[key];
|
|
36
|
+
if (isObject(existing) && isObject(value)) {
|
|
37
|
+
result[key] = mergeJsonDefaults(existing as JSONObject, value as JSONObject);
|
|
38
|
+
continue;
|
|
39
|
+
}
|
|
40
|
+
result[key] = value as JSONValue;
|
|
41
|
+
}
|
|
42
|
+
|
|
43
|
+
return result;
|
|
44
|
+
}
|
|
45
|
+
|
|
46
|
+
export function mergeSmartAiProviderOptions(
|
|
47
|
+
defaults?: TSmartAiProviderOptions,
|
|
48
|
+
overrides?: TSmartAiProviderOptions,
|
|
49
|
+
): TSmartAiProviderOptions | undefined {
|
|
50
|
+
if (!defaults) return overrides;
|
|
51
|
+
if (!overrides) return defaults;
|
|
52
|
+
return mergeJsonDefaults(defaults as JSONObject, overrides as JSONObject) as TSmartAiProviderOptions;
|
|
53
|
+
}
|
|
54
|
+
|
|
55
|
+
function cacheOptionsFromSetting(cache: TSmartAiCacheSetting | undefined): ISmartAiCacheOptions | undefined {
|
|
56
|
+
if (cache === false) return undefined;
|
|
57
|
+
if (cache === undefined || cache === true || cache === 'auto') return {};
|
|
58
|
+
return cache;
|
|
59
|
+
}
|
|
60
|
+
|
|
61
|
+
export function resolveSmartAiCacheProvider(provider?: string, modelId?: string): TSmartAiMessageCacheProvider | undefined {
|
|
62
|
+
const providerLower = provider?.toLowerCase() ?? '';
|
|
63
|
+
const modelLower = modelId?.toLowerCase() ?? '';
|
|
64
|
+
|
|
65
|
+
if (providerLower.includes('openrouter')) return 'openrouter';
|
|
66
|
+
if (providerLower.includes('bedrock')) return 'bedrock';
|
|
67
|
+
if (providerLower.includes('copilot')) return 'copilot';
|
|
68
|
+
if (providerLower.includes('alibaba')) return 'alibaba';
|
|
69
|
+
if (providerLower.includes('openai-compatible') || providerLower.includes('openaicompatible')) {
|
|
70
|
+
return 'openaiCompatible';
|
|
71
|
+
}
|
|
72
|
+
if (providerLower.includes('anthropic')) return 'anthropic';
|
|
73
|
+
if (modelLower.includes('claude') || modelLower.includes('anthropic')) return 'anthropic';
|
|
74
|
+
|
|
75
|
+
return undefined;
|
|
76
|
+
}
|
|
77
|
+
|
|
78
|
+
export function getSmartAiMessageCacheProviderOptions(
|
|
79
|
+
provider: TSmartAiMessageCacheProvider,
|
|
80
|
+
options: ISmartAiCacheOptions = {},
|
|
81
|
+
): TSmartAiProviderOptions {
|
|
82
|
+
const anthropicCacheControl: JSONObject = {
|
|
83
|
+
type: 'ephemeral',
|
|
84
|
+
...(options.retention === '1h' ? { ttl: '1h' } : {}),
|
|
85
|
+
};
|
|
86
|
+
|
|
87
|
+
const providerOptions: Record<TSmartAiMessageCacheProvider, JSONObject> = {
|
|
88
|
+
anthropic: {
|
|
89
|
+
anthropic: {
|
|
90
|
+
cacheControl: anthropicCacheControl,
|
|
91
|
+
},
|
|
92
|
+
},
|
|
93
|
+
openrouter: {
|
|
94
|
+
openrouter: {
|
|
95
|
+
cacheControl: { type: 'ephemeral' },
|
|
96
|
+
},
|
|
97
|
+
},
|
|
98
|
+
bedrock: {
|
|
99
|
+
bedrock: {
|
|
100
|
+
cachePoint: { type: 'default' },
|
|
101
|
+
},
|
|
102
|
+
},
|
|
103
|
+
openaiCompatible: {
|
|
104
|
+
openaiCompatible: {
|
|
105
|
+
cache_control: { type: 'ephemeral' },
|
|
106
|
+
},
|
|
107
|
+
},
|
|
108
|
+
copilot: {
|
|
109
|
+
copilot: {
|
|
110
|
+
copilot_cache_control: { type: 'ephemeral' },
|
|
111
|
+
},
|
|
112
|
+
},
|
|
113
|
+
alibaba: {
|
|
114
|
+
alibaba: {
|
|
115
|
+
cacheControl: { type: 'ephemeral' },
|
|
116
|
+
},
|
|
117
|
+
},
|
|
118
|
+
};
|
|
119
|
+
|
|
120
|
+
return providerOptions[provider] as TSmartAiProviderOptions;
|
|
121
|
+
}
|
|
122
|
+
|
|
123
|
+
function shouldUseMessageLevelOptions(provider: TSmartAiMessageCacheProvider): boolean {
|
|
124
|
+
return provider === 'anthropic' || provider === 'bedrock';
|
|
125
|
+
}
|
|
126
|
+
|
|
127
|
+
function applyProviderOptionsDefaults<T extends { providerOptions?: TSmartAiProviderOptions }>(
|
|
128
|
+
item: T,
|
|
129
|
+
defaults: TSmartAiProviderOptions,
|
|
130
|
+
): T {
|
|
131
|
+
return {
|
|
132
|
+
...item,
|
|
133
|
+
providerOptions: mergeSmartAiProviderOptions(defaults, item.providerOptions),
|
|
134
|
+
};
|
|
135
|
+
}
|
|
136
|
+
|
|
137
|
+
function isToolApprovalPart(part: unknown): boolean {
|
|
138
|
+
if (!isObject(part)) return false;
|
|
139
|
+
return part.type === 'tool-approval-request' || part.type === 'tool-approval-response';
|
|
140
|
+
}
|
|
141
|
+
|
|
142
|
+
function applyCacheToMessage(
|
|
143
|
+
message: LanguageModelV3Prompt[number],
|
|
144
|
+
provider: TSmartAiMessageCacheProvider,
|
|
145
|
+
options: ISmartAiCacheOptions,
|
|
146
|
+
): LanguageModelV3Prompt[number] {
|
|
147
|
+
const providerOptions = getSmartAiMessageCacheProviderOptions(provider, options);
|
|
148
|
+
const content = message.content;
|
|
149
|
+
|
|
150
|
+
if (!shouldUseMessageLevelOptions(provider) && Array.isArray(content) && content.length > 0) {
|
|
151
|
+
const lastIndex = content.length - 1;
|
|
152
|
+
const lastPart = content[lastIndex];
|
|
153
|
+
if (!isToolApprovalPart(lastPart)) {
|
|
154
|
+
const messageWithArrayContent = message as Extract<LanguageModelV3Prompt[number], { content: unknown[] }>;
|
|
155
|
+
return {
|
|
156
|
+
...messageWithArrayContent,
|
|
157
|
+
content: content.map((part, index) =>
|
|
158
|
+
index === lastIndex ? applyProviderOptionsDefaults(part, providerOptions) : part,
|
|
159
|
+
) as typeof messageWithArrayContent.content,
|
|
160
|
+
} as LanguageModelV3Prompt[number];
|
|
161
|
+
}
|
|
162
|
+
}
|
|
163
|
+
|
|
164
|
+
return applyProviderOptionsDefaults(message, providerOptions);
|
|
165
|
+
}
|
|
166
|
+
|
|
167
|
+
export function applySmartAiPromptCaching(
|
|
168
|
+
prompt: LanguageModelV3Prompt,
|
|
169
|
+
options: ISmartAiCacheOptions = {},
|
|
170
|
+
): LanguageModelV3Prompt {
|
|
171
|
+
const provider = options.provider ?? 'anthropic';
|
|
172
|
+
const targetIndexes = new Set<number>();
|
|
173
|
+
const nonSystemIndexes: number[] = [];
|
|
174
|
+
let systemCount = 0;
|
|
175
|
+
|
|
176
|
+
for (let i = 0; i < prompt.length; i++) {
|
|
177
|
+
const message = prompt[i];
|
|
178
|
+
if (message.role === 'system') {
|
|
179
|
+
if (systemCount < 2) targetIndexes.add(i);
|
|
180
|
+
systemCount++;
|
|
181
|
+
continue;
|
|
182
|
+
}
|
|
183
|
+
nonSystemIndexes.push(i);
|
|
184
|
+
}
|
|
185
|
+
|
|
186
|
+
for (const index of nonSystemIndexes.slice(-2)) {
|
|
187
|
+
targetIndexes.add(index);
|
|
188
|
+
}
|
|
189
|
+
|
|
190
|
+
if (targetIndexes.size === 0) return prompt;
|
|
191
|
+
|
|
192
|
+
return prompt.map((message, index) =>
|
|
193
|
+
targetIndexes.has(index) ? applyCacheToMessage(message, provider, options) : message,
|
|
194
|
+
) as LanguageModelV3Prompt;
|
|
195
|
+
}
|
|
196
|
+
|
|
197
|
+
export function createSmartAiCachingMiddleware(options: ISmartAiCacheOptions = {}): LanguageModelV3Middleware {
|
|
198
|
+
return {
|
|
199
|
+
specificationVersion: 'v3',
|
|
200
|
+
transformParams: async ({ params }) => ({
|
|
201
|
+
...params,
|
|
202
|
+
prompt: applySmartAiPromptCaching(params.prompt, options),
|
|
203
|
+
}),
|
|
204
|
+
};
|
|
205
|
+
}
|
|
206
|
+
|
|
207
|
+
function isOpenAiProvider(provider?: string): boolean {
|
|
208
|
+
const providerLower = provider?.toLowerCase() ?? '';
|
|
209
|
+
return providerLower === 'openai' || providerLower.startsWith('openai.') || providerLower.includes('@ai-sdk/openai');
|
|
210
|
+
}
|
|
211
|
+
|
|
212
|
+
export function getSmartAiCacheProviderOptions(input: {
|
|
213
|
+
provider?: string;
|
|
214
|
+
modelId?: string;
|
|
215
|
+
cache?: TSmartAiCacheSetting;
|
|
216
|
+
sessionId?: string;
|
|
217
|
+
}): TSmartAiProviderOptions | undefined {
|
|
218
|
+
const cacheOptions = cacheOptionsFromSetting(input.cache);
|
|
219
|
+
if (!cacheOptions) return undefined;
|
|
220
|
+
|
|
221
|
+
if (isOpenAiProvider(input.provider)) {
|
|
222
|
+
const key = cacheOptions.key ?? input.sessionId;
|
|
223
|
+
return {
|
|
224
|
+
openai: {
|
|
225
|
+
store: false,
|
|
226
|
+
...(key ? { promptCacheKey: key } : {}),
|
|
227
|
+
...(cacheOptions.retention === '24h' || cacheOptions.retention === 'in_memory'
|
|
228
|
+
? { promptCacheRetention: cacheOptions.retention }
|
|
229
|
+
: key
|
|
230
|
+
? { promptCacheRetention: 'in_memory' }
|
|
231
|
+
: {}),
|
|
232
|
+
},
|
|
233
|
+
};
|
|
234
|
+
}
|
|
235
|
+
|
|
236
|
+
return undefined;
|
|
237
|
+
}
|
|
238
|
+
|
|
239
|
+
export function applySmartAiCacheProviderOptions(input: {
|
|
240
|
+
provider?: string;
|
|
241
|
+
modelId?: string;
|
|
242
|
+
providerOptions?: TSmartAiProviderOptions;
|
|
243
|
+
cache?: TSmartAiCacheSetting;
|
|
244
|
+
sessionId?: string;
|
|
245
|
+
}): TSmartAiProviderOptions | undefined {
|
|
246
|
+
return mergeSmartAiProviderOptions(
|
|
247
|
+
getSmartAiCacheProviderOptions(input),
|
|
248
|
+
input.providerOptions,
|
|
249
|
+
);
|
|
250
|
+
}
|
|
@@ -2,6 +2,7 @@ import * as plugins from './plugins.js';
|
|
|
2
2
|
import type { ISmartAiModelSetup, ISmartAiOptions, LanguageModelV3 } from './smartai.interfaces.js';
|
|
3
3
|
import { createOllamaModel } from './smartai.provider.ollama.js';
|
|
4
4
|
import { createAnthropicCachingMiddleware } from './smartai.middleware.anthropic.js';
|
|
5
|
+
import { createOpenAiChatGptProviderSettings } from './smartai.auth.openai.js';
|
|
5
6
|
|
|
6
7
|
/**
|
|
7
8
|
* Returns a LanguageModelV3 for the given provider and model.
|
|
@@ -16,11 +17,17 @@ export function getModel(options: ISmartAiOptions): LanguageModelV3 {
|
|
|
16
17
|
if (options.promptCaching === false) return base;
|
|
17
18
|
return plugins.wrapLanguageModel({
|
|
18
19
|
model: base,
|
|
19
|
-
middleware: createAnthropicCachingMiddleware(
|
|
20
|
+
middleware: createAnthropicCachingMiddleware(
|
|
21
|
+
typeof options.promptCaching === 'object' ? options.promptCaching : undefined,
|
|
22
|
+
),
|
|
20
23
|
}) as unknown as LanguageModelV3;
|
|
21
24
|
}
|
|
22
25
|
case 'openai': {
|
|
23
|
-
const p = plugins.createOpenAI(
|
|
26
|
+
const p = plugins.createOpenAI(
|
|
27
|
+
options.openAiChatGptAuth
|
|
28
|
+
? createOpenAiChatGptProviderSettings(options.openAiChatGptAuth)
|
|
29
|
+
: { apiKey: options.apiKey },
|
|
30
|
+
);
|
|
24
31
|
return p(options.model) as LanguageModelV3;
|
|
25
32
|
}
|
|
26
33
|
case 'google': {
|
package/ts/smartai.interfaces.ts
CHANGED
|
@@ -1,4 +1,5 @@
|
|
|
1
|
-
import type { JSONObject, JSONValue, LanguageModelV3 } from '@ai-sdk/provider';
|
|
1
|
+
import type { JSONObject, JSONValue, LanguageModelV3, LanguageModelV3Prompt } from '@ai-sdk/provider';
|
|
2
|
+
import type { ISmartAiCacheOptions } from './smartai.cache.js';
|
|
2
3
|
|
|
3
4
|
export type TProvider =
|
|
4
5
|
| 'anthropic'
|
|
@@ -14,6 +15,53 @@ export type TOpenAiReasoningEffort = 'none' | 'minimal' | 'low' | 'medium' | 'hi
|
|
|
14
15
|
|
|
15
16
|
export type TOpenAiTextVerbosity = 'low' | 'medium' | 'high';
|
|
16
17
|
|
|
18
|
+
export interface IOpenAiChatGptTokenInfo {
|
|
19
|
+
email?: string;
|
|
20
|
+
chatgptPlanType?: string;
|
|
21
|
+
chatgptUserId?: string;
|
|
22
|
+
chatgptAccountId?: string;
|
|
23
|
+
chatgptAccountIsFedramp: boolean;
|
|
24
|
+
expiresAt?: string;
|
|
25
|
+
rawJwt: string;
|
|
26
|
+
}
|
|
27
|
+
|
|
28
|
+
export interface IOpenAiChatGptAuthCredentials {
|
|
29
|
+
accessToken: string;
|
|
30
|
+
refreshToken?: string;
|
|
31
|
+
idToken?: string;
|
|
32
|
+
accountId?: string;
|
|
33
|
+
tokenInfo?: IOpenAiChatGptTokenInfo;
|
|
34
|
+
baseUrl?: string;
|
|
35
|
+
originator?: string;
|
|
36
|
+
}
|
|
37
|
+
|
|
38
|
+
export interface IOpenAiChatGptTokenData extends IOpenAiChatGptAuthCredentials {
|
|
39
|
+
refreshToken: string;
|
|
40
|
+
tokenInfo: IOpenAiChatGptTokenInfo;
|
|
41
|
+
}
|
|
42
|
+
|
|
43
|
+
export interface IOpenAiChatGptDeviceCode {
|
|
44
|
+
verificationUrl: string;
|
|
45
|
+
userCode: string;
|
|
46
|
+
deviceAuthId: string;
|
|
47
|
+
intervalSeconds: number;
|
|
48
|
+
}
|
|
49
|
+
|
|
50
|
+
export interface IOpenAiChatGptAuthOptions {
|
|
51
|
+
issuer?: string;
|
|
52
|
+
clientId?: string;
|
|
53
|
+
fetch?: typeof fetch;
|
|
54
|
+
}
|
|
55
|
+
|
|
56
|
+
export interface IOpenAiChatGptDeviceCodePollOptions extends IOpenAiChatGptAuthOptions {
|
|
57
|
+
timeoutMs?: number;
|
|
58
|
+
sleep?: (ms: number) => Promise<void>;
|
|
59
|
+
}
|
|
60
|
+
|
|
61
|
+
export interface IOpenAiChatGptCompleteDeviceCodeOptions extends IOpenAiChatGptDeviceCodePollOptions {
|
|
62
|
+
forcedChatGptWorkspaceId?: string;
|
|
63
|
+
}
|
|
64
|
+
|
|
17
65
|
export interface IOpenAiProviderOptions extends JSONObject {
|
|
18
66
|
conversation?: string | null;
|
|
19
67
|
include?: string[] | null;
|
|
@@ -55,6 +103,11 @@ export interface ISmartAiOptions {
|
|
|
55
103
|
provider: TProvider;
|
|
56
104
|
model: string;
|
|
57
105
|
apiKey?: string;
|
|
106
|
+
/**
|
|
107
|
+
* OpenAI ChatGPT/Codex subscription credentials from the device-code auth flow.
|
|
108
|
+
* Only used when provider === 'openai'.
|
|
109
|
+
*/
|
|
110
|
+
openAiChatGptAuth?: IOpenAiChatGptAuthCredentials;
|
|
58
111
|
/**
|
|
59
112
|
* Provider-specific AI SDK generation options.
|
|
60
113
|
* Pass this to generateText()/streamText() alongside the model.
|
|
@@ -71,7 +124,7 @@ export interface ISmartAiOptions {
|
|
|
71
124
|
* Enable Anthropic prompt caching on system + recent messages.
|
|
72
125
|
* Only used when provider === 'anthropic'. Default: true.
|
|
73
126
|
*/
|
|
74
|
-
promptCaching?: boolean;
|
|
127
|
+
promptCaching?: boolean | ISmartAiCacheOptions;
|
|
75
128
|
}
|
|
76
129
|
|
|
77
130
|
/**
|
|
@@ -96,4 +149,4 @@ export interface IOllamaModelOptions {
|
|
|
96
149
|
think?: boolean;
|
|
97
150
|
}
|
|
98
151
|
|
|
99
|
-
export type { LanguageModelV3 };
|
|
152
|
+
export type { LanguageModelV3, LanguageModelV3Prompt };
|
|
@@ -1,38 +1,12 @@
|
|
|
1
|
-
import type { LanguageModelV3Middleware
|
|
1
|
+
import type { LanguageModelV3Middleware } from '@ai-sdk/provider';
|
|
2
|
+
import { createSmartAiCachingMiddleware } from './smartai.cache.js';
|
|
3
|
+
import type { ISmartAiCacheOptions } from './smartai.cache.js';
|
|
2
4
|
|
|
3
5
|
/**
|
|
4
6
|
* Creates middleware that adds Anthropic prompt caching directives.
|
|
5
7
|
* Marks the last system message and last user message with ephemeral cache control,
|
|
6
8
|
* reducing input token cost and latency on repeated calls.
|
|
7
9
|
*/
|
|
8
|
-
export function createAnthropicCachingMiddleware(): LanguageModelV3Middleware {
|
|
9
|
-
return {
|
|
10
|
-
specificationVersion: 'v3',
|
|
11
|
-
transformParams: async ({ params }) => {
|
|
12
|
-
const messages = [...params.prompt] as Array<Record<string, unknown>>;
|
|
13
|
-
|
|
14
|
-
// Find the last system message and last user message
|
|
15
|
-
let lastSystemIdx = -1;
|
|
16
|
-
let lastUserIdx = -1;
|
|
17
|
-
for (let i = 0; i < messages.length; i++) {
|
|
18
|
-
if (messages[i].role === 'system') lastSystemIdx = i;
|
|
19
|
-
if (messages[i].role === 'user') lastUserIdx = i;
|
|
20
|
-
}
|
|
21
|
-
|
|
22
|
-
const targets = [lastSystemIdx, lastUserIdx].filter(i => i >= 0);
|
|
23
|
-
for (const idx of targets) {
|
|
24
|
-
const msg = { ...messages[idx] };
|
|
25
|
-
msg.providerOptions = {
|
|
26
|
-
...(msg.providerOptions as Record<string, unknown> || {}),
|
|
27
|
-
anthropic: {
|
|
28
|
-
...((msg.providerOptions as Record<string, unknown>)?.anthropic as Record<string, unknown> || {}),
|
|
29
|
-
cacheControl: { type: 'ephemeral' },
|
|
30
|
-
},
|
|
31
|
-
};
|
|
32
|
-
messages[idx] = msg;
|
|
33
|
-
}
|
|
34
|
-
|
|
35
|
-
return { ...params, prompt: messages as unknown as LanguageModelV3Prompt };
|
|
36
|
-
},
|
|
37
|
-
};
|
|
10
|
+
export function createAnthropicCachingMiddleware(options: ISmartAiCacheOptions = {}): LanguageModelV3Middleware {
|
|
11
|
+
return createSmartAiCachingMiddleware({ ...options, provider: 'anthropic' });
|
|
38
12
|
}
|