utilitas 1999.1.41 → 1999.1.43
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +0 -1
- package/dist/utilitas.lite.mjs +1 -1
- package/dist/utilitas.lite.mjs.map +1 -1
- package/lib/alan.mjs +25 -31
- package/lib/manifest.mjs +1 -1
- package/package.json +1 -1
package/lib/alan.mjs
CHANGED
|
@@ -54,22 +54,22 @@ const [
|
|
|
54
54
|
];
|
|
55
55
|
|
|
56
56
|
const [
|
|
57
|
-
OPENAI, GEMINI, OPENAI_TRAINING, OLLAMA,
|
|
58
|
-
|
|
59
|
-
|
|
60
|
-
|
|
61
|
-
|
|
62
|
-
|
|
63
|
-
|
|
64
|
-
|
|
65
|
-
|
|
66
|
-
|
|
67
|
-
|
|
68
|
-
|
|
57
|
+
OPENAI, GEMINI, OPENAI_TRAINING, OLLAMA, GPT_4O, GPT_O1, GPT_O3_MINI,
|
|
58
|
+
GEMINI_20_FLASH, NOVA, DEEPSEEK_R1, MD_CODE, TEXT_EMBEDDING_3_SMALL,
|
|
59
|
+
TEXT_EMBEDDING_3_LARGE, CLOUD_37_SONNET, AUDIO, WAV, ATTACHMENTS, CHAT,
|
|
60
|
+
OPENAI_VOICE, MEDIUM, LOW, HIGH, GPT_REASONING_EFFORT, THINK, THINK_STR,
|
|
61
|
+
THINK_END, AZURE, TOOLS_STR, TOOLS_END, TOOLS, TEXT, THINKING, OK, FUNC,
|
|
62
|
+
GPT_45, REDACTED_THINKING, GEMMA_3_27B, AZURE_OPENAI, ANTHROPIC,
|
|
63
|
+
VERTEX_ANTHROPIC, GEMMA327B, v8k, ais, MAX_TOOL_RECURSION, LOG, name, user,
|
|
64
|
+
system, assistant, MODEL, JSON_OBJECT, TOOL, silent, GEMINI_EMBEDDING_M,
|
|
65
|
+
INVALID_FILE, tokenSafeRatio, GPT_QUERY_LIMIT, CONTENT_IS_REQUIRED,
|
|
66
|
+
OPENAI_HI_RES_SIZE, k, kT, m, minute, hour, gb, trimTailing, EBD,
|
|
67
|
+
GEMINI_20_FLASH_EXP, IMAGE, JINA, JINA_DEEPSEARCH, JINA_CLIP, VERTEX,
|
|
68
|
+
GEMINI_25_PRO, SILICONFLOW, SF_DEEPSEEK_V3,
|
|
69
69
|
] = [
|
|
70
|
-
'OpenAI', 'Gemini', 'OPENAI_TRAINING', 'Ollama', 'gpt-4o
|
|
71
|
-
'
|
|
72
|
-
'
|
|
70
|
+
'OpenAI', 'Gemini', 'OPENAI_TRAINING', 'Ollama', 'gpt-4o', 'o1',
|
|
71
|
+
'o3-mini', 'gemini-2.0-flash', 'nova', 'deepseek-r1', '```',
|
|
72
|
+
'text-embedding-3-small', 'text-embedding-3-large',
|
|
73
73
|
'claude-3-7-sonnet@20250219', 'audio', 'wav', '[ATTACHMENTS]', 'CHAT',
|
|
74
74
|
'OPENAI_VOICE', 'medium', 'low', 'high', 'medium', 'think', '<think>',
|
|
75
75
|
'</think>', 'AZURE', '<tools>', '</tools>', 'tools', 'text', 'thinking',
|
|
@@ -108,13 +108,12 @@ const AzureOpenAI = async opts => new (await libOpenAi(opts)).AzureOpenAI(opts);
|
|
|
108
108
|
const OPENAI_S1 = { contextWindow: kT(128), maxOutputTokens: k(16) };
|
|
109
109
|
const OPENAI_S2 = { contextWindow: kT(200), maxOutputTokens: kT(100) };
|
|
110
110
|
const OPENAI_EBD = { ...EBD, maxInputTokens: k(8) - 1 };
|
|
111
|
-
const OPENAI_AUDIO_TYPES = { supportedAudioTypes: [wav] };
|
|
112
|
-
const GPT_4O_AUDIO = { ...OPENAI_AUDIO_TYPES, audio: 'gpt-4o-audio-preview' };
|
|
113
111
|
|
|
114
112
|
const OPENAI_RULES = {
|
|
115
113
|
imageCostTokens: ~~(OPENAI_HI_RES_SIZE / (512 * 512) * 170 + 85),
|
|
116
114
|
maxFileSize: m(20), maxImageSize: OPENAI_HI_RES_SIZE,
|
|
117
115
|
supportedMimeTypes: [png, jpeg, gif, webp],
|
|
116
|
+
supportedAudioTypes: [wav], audio: 'gpt-4o-audio-preview',
|
|
118
117
|
json: true, tools: true, vision: true, defaultProvider: OPENAI,
|
|
119
118
|
};
|
|
120
119
|
|
|
@@ -132,19 +131,12 @@ const GEMINI_RULES = {
|
|
|
132
131
|
// https://platform.openai.com/docs/models
|
|
133
132
|
// https://cloud.google.com/vertex-ai/docs/generative-ai/learn/models
|
|
134
133
|
const MODELS = {
|
|
135
|
-
[GPT_4O]: { ...
|
|
136
|
-
[GPT_45]: { ...
|
|
137
|
-
[
|
|
138
|
-
...OPENAI_S1, ...OPENAI_RULES, ...OPENAI_AUDIO_TYPES,
|
|
139
|
-
audio: 'gpt-4o-mini-audio-preview', fast: true,
|
|
140
|
-
},
|
|
141
|
-
[GPT_O1]: {
|
|
142
|
-
...OPENAI_S2, ...OPENAI_RULES, ...GPT_4O_AUDIO,
|
|
143
|
-
reasoning: true,
|
|
144
|
-
},
|
|
134
|
+
[GPT_4O]: { ...OPENAI_RULES, ...OPENAI_S1 },
|
|
135
|
+
[GPT_45]: { ...OPENAI_RULES, ...OPENAI_S1 },
|
|
136
|
+
[GPT_O1]: { ...OPENAI_RULES, ...OPENAI_S2, reasoning: true },
|
|
145
137
|
[GPT_O3_MINI]: {
|
|
146
|
-
...
|
|
147
|
-
fast: true, reasoning: true,
|
|
138
|
+
...OPENAI_RULES, ...OPENAI_S2,
|
|
139
|
+
fast: true, reasoning: true, vision: GPT_O1,
|
|
148
140
|
},
|
|
149
141
|
[GEMINI_20_FLASH]: {
|
|
150
142
|
...GEMINI_RULES, contextWindow: m(1), maxOutputTokens: k(8),
|
|
@@ -228,7 +220,7 @@ const DEFAULT_MODELS = {
|
|
|
228
220
|
[JINA]: JINA_DEEPSEARCH,
|
|
229
221
|
[OLLAMA]: GEMMA327B,
|
|
230
222
|
[OPENAI_VOICE]: NOVA,
|
|
231
|
-
[OPENAI_TRAINING]:
|
|
223
|
+
[OPENAI_TRAINING]: GPT_4O, // https://platform.openai.com/docs/guides/fine-tuning
|
|
232
224
|
};
|
|
233
225
|
|
|
234
226
|
const DEFAULT_EMBEDDING = {
|
|
@@ -590,6 +582,9 @@ const buildGptMessage = (content, options) => {
|
|
|
590
582
|
const attachments = (options?.attachments || []).map(x => {
|
|
591
583
|
assert(MODELS[options?.model], 'Model is required.');
|
|
592
584
|
if (MODELS[options.model]?.supportedMimeTypes?.includes?.(x.mime_type)) {
|
|
585
|
+
if (String.isString(MODELS[options.model]?.vision)) {
|
|
586
|
+
alterModel = MODELS[options.model]?.vision;
|
|
587
|
+
}
|
|
593
588
|
return { type: 'image_url', image_url: { url: x.url, detail: 'high' } };
|
|
594
589
|
} else if (MODELS[options.model]?.supportedAudioTypes?.includes?.(x.mime_type)) {
|
|
595
590
|
alterModel = selectGptAudioModel(options);
|
|
@@ -1541,7 +1536,6 @@ export {
|
|
|
1541
1536
|
FUNCTION,
|
|
1542
1537
|
GEMINI_20_FLASH,
|
|
1543
1538
|
GPT_45,
|
|
1544
|
-
GPT_4O_MINI,
|
|
1545
1539
|
GPT_4O,
|
|
1546
1540
|
GPT_O1,
|
|
1547
1541
|
GPT_O3_MINI,
|
package/lib/manifest.mjs
CHANGED