omnikey-cli 1.0.38 → 1.0.39
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/backend-dist/ai-client.js +7 -7
- package/backend-dist/featureRoutes.js +2 -2
- package/backend-dist/index.js +2 -2
- package/dist/onboard.js +2 -2
- package/package.json +3 -3
- package/src/onboard.ts +2 -2
|
@@ -16,9 +16,9 @@ const config_1 = require("./config");
|
|
|
16
16
|
// Default model mapping
|
|
17
17
|
// ---------------------------------------------------------------------------
|
|
18
18
|
const DEFAULT_MODELS = {
|
|
19
|
-
openai: { fast: 'gpt-4o-mini', smart: 'gpt-5.
|
|
19
|
+
openai: { fast: 'gpt-4o-mini', smart: 'gpt-5.5' },
|
|
20
20
|
gemini: { fast: 'gemini-2.5-flash', smart: 'gemini-2.5-pro' },
|
|
21
|
-
anthropic: { fast: 'claude-haiku-4-5-20251001', smart: 'claude-
|
|
21
|
+
anthropic: { fast: 'claude-haiku-4-5-20251001', smart: 'claude-opus-4-7' },
|
|
22
22
|
};
|
|
23
23
|
function getDefaultModel(provider, tier) {
|
|
24
24
|
return DEFAULT_MODELS[provider][tier];
|
|
@@ -29,7 +29,7 @@ function getDefaultModel(provider, tier) {
|
|
|
29
29
|
* - anthropic: hard API-enforced string limit of 10,485,760 chars; we stay
|
|
30
30
|
* just below it with a small safety buffer.
|
|
31
31
|
* - openai: no documented per-string limit; bounded by the context window
|
|
32
|
-
* (~272K tokens for GPT-5.
|
|
32
|
+
* (~272K tokens for GPT-5.5 ≈ ~1M chars). Use the history cap.
|
|
33
33
|
* - gemini: no documented per-string limit; bounded by the 1M-token
|
|
34
34
|
* context window (~4M chars). Use the history cap.
|
|
35
35
|
*/
|
|
@@ -43,9 +43,9 @@ const MAX_MESSAGE_CONTENT_LENGTH_BY_PROVIDER = {
|
|
|
43
43
|
* history, derived from each provider's context-window size minus headroom
|
|
44
44
|
* for the system prompt and max output tokens.
|
|
45
45
|
*
|
|
46
|
-
* - anthropic: Claude
|
|
46
|
+
* - anthropic: Claude Opus 4.7 — 1M token ctx, 64K max output
|
|
47
47
|
* ≈ (1,000,000 - 64,000 - 10,000) tokens × 4 chars ≈ 3.7M chars
|
|
48
|
-
* - openai: GPT-5.
|
|
48
|
+
* - openai: GPT-5.5 — ~272K token ctx, ~32K max output
|
|
49
49
|
* ≈ (272,000 - 32,000 - 5,000) tokens × 4 chars ≈ 940K chars
|
|
50
50
|
* - gemini: Gemini 2.5 Pro — 1M token ctx, ~32K max output
|
|
51
51
|
* ≈ (1,000,000 - 32,000 - 10,000) tokens × 4 chars ≈ 3.8M chars
|
|
@@ -75,7 +75,7 @@ class OpenAIAdapter {
|
|
|
75
75
|
model,
|
|
76
76
|
messages: oaiMessages,
|
|
77
77
|
tools: tools?.length ? tools : undefined,
|
|
78
|
-
temperature: options.temperature ?? 0.2,
|
|
78
|
+
temperature: model === 'gpt-5.5' ? 1 : (options.temperature ?? 0.2),
|
|
79
79
|
max_tokens: options.maxTokens,
|
|
80
80
|
});
|
|
81
81
|
const choice = completion.choices[0];
|
|
@@ -175,7 +175,7 @@ class AnthropicAdapter {
|
|
|
175
175
|
...(system ? { system } : {}),
|
|
176
176
|
messages: anthropicMessages,
|
|
177
177
|
...(tools?.length ? { tools } : {}),
|
|
178
|
-
temperature: options.temperature ?? 0.2,
|
|
178
|
+
...(model === 'claude-opus-4-7' ? {} : { temperature: options.temperature ?? 0.2 }),
|
|
179
179
|
});
|
|
180
180
|
const textContent = response.content
|
|
181
181
|
.filter((b) => b.type === 'text')
|
|
@@ -58,9 +58,9 @@ async function getPromptForCommand(logger, cmd, subscription) {
|
|
|
58
58
|
function getModelForCommand(cmd) {
|
|
59
59
|
const tier = cmd === 'task' ? 'smart' : 'fast';
|
|
60
60
|
const models = {
|
|
61
|
-
openai: { fast: 'gpt-4o-mini', smart: 'gpt-5.
|
|
61
|
+
openai: { fast: 'gpt-4o-mini', smart: 'gpt-5.5' },
|
|
62
62
|
gemini: { fast: 'gemini-2.5-flash', smart: 'gemini-2.5-pro' },
|
|
63
|
-
anthropic: { fast: 'claude-haiku-4-5-20251001', smart: 'claude-
|
|
63
|
+
anthropic: { fast: 'claude-haiku-4-5-20251001', smart: 'claude-opus-4-7' },
|
|
64
64
|
};
|
|
65
65
|
return models[config_1.config.aiProvider]?.[tier] ?? 'gpt-4o-mini';
|
|
66
66
|
}
|
package/backend-dist/index.js
CHANGED
|
@@ -74,8 +74,8 @@ app.get('/macos/appcast', (req, res) => {
|
|
|
74
74
|
const appcastUrl = `${baseUrl}/macos/appcast`;
|
|
75
75
|
// These should match the values embedded into the macOS app
|
|
76
76
|
// Info.plist in macOS/build_release_dmg.sh.
|
|
77
|
-
const bundleVersion = '
|
|
78
|
-
const shortVersion = '1.0.
|
|
77
|
+
const bundleVersion = '26';
|
|
78
|
+
const shortVersion = '1.0.25';
|
|
79
79
|
const xml = `<?xml version="1.0" encoding="utf-8"?>
|
|
80
80
|
<rss version="2.0"
|
|
81
81
|
xmlns:sparkle="http://www.andymatuschak.org/xml-namespaces/sparkle"
|
package/dist/onboard.js
CHANGED
|
@@ -9,8 +9,8 @@ const fs_1 = __importDefault(require("fs"));
|
|
|
9
9
|
const path_1 = __importDefault(require("path"));
|
|
10
10
|
const utils_1 = require("./utils");
|
|
11
11
|
const AI_PROVIDERS = [
|
|
12
|
-
{ name: 'OpenAI (gpt-4o-mini / gpt-5.
|
|
13
|
-
{ name: 'Anthropic — Claude (claude-haiku / claude-
|
|
12
|
+
{ name: 'OpenAI (gpt-4o-mini / gpt-5.5)', value: 'openai' },
|
|
13
|
+
{ name: 'Anthropic — Claude (claude-haiku / claude-opus)', value: 'anthropic' },
|
|
14
14
|
{ name: 'Google Gemini (gemini-2.5-flash / gemini-2.5-pro)', value: 'gemini' },
|
|
15
15
|
];
|
|
16
16
|
const SEARCH_PROVIDERS = [
|
package/package.json
CHANGED
|
@@ -4,7 +4,7 @@
|
|
|
4
4
|
"access": "public",
|
|
5
5
|
"registry": "https://registry.npmjs.org/"
|
|
6
6
|
},
|
|
7
|
-
"version": "1.0.
|
|
7
|
+
"version": "1.0.39",
|
|
8
8
|
"description": "CLI for onboarding users to Omnikey AI and configuring OPENAI_API_KEY. Use Yarn for install/build.",
|
|
9
9
|
"engines": {
|
|
10
10
|
"node": ">=14.0.0",
|
|
@@ -27,7 +27,7 @@
|
|
|
27
27
|
"author": "Gurinder Rawala",
|
|
28
28
|
"license": "MIT",
|
|
29
29
|
"dependencies": {
|
|
30
|
-
"@anthropic-ai/sdk": "^0.
|
|
30
|
+
"@anthropic-ai/sdk": "^0.96.0",
|
|
31
31
|
"@google-cloud/storage": "^7.19.0",
|
|
32
32
|
"@google/genai": "^1.46.0",
|
|
33
33
|
"axios": "^1.13.5",
|
|
@@ -38,7 +38,7 @@
|
|
|
38
38
|
"express": "^4.21.2",
|
|
39
39
|
"inquirer": "^9.0.0",
|
|
40
40
|
"jsonwebtoken": "^9.0.3",
|
|
41
|
-
"openai": "^6.
|
|
41
|
+
"openai": "^6.37.0",
|
|
42
42
|
"pg": "^8.18.0",
|
|
43
43
|
"pg-hstore": "^2.3.4",
|
|
44
44
|
"playwright-core": "^1.50.0",
|
package/src/onboard.ts
CHANGED
|
@@ -4,8 +4,8 @@ import path from 'path';
|
|
|
4
4
|
import { getConfigDir, getConfigPath } from './utils';
|
|
5
5
|
|
|
6
6
|
const AI_PROVIDERS = [
|
|
7
|
-
{ name: 'OpenAI (gpt-4o-mini / gpt-5.
|
|
8
|
-
{ name: 'Anthropic — Claude (claude-haiku / claude-
|
|
7
|
+
{ name: 'OpenAI (gpt-4o-mini / gpt-5.5)', value: 'openai' },
|
|
8
|
+
{ name: 'Anthropic — Claude (claude-haiku / claude-opus)', value: 'anthropic' },
|
|
9
9
|
{ name: 'Google Gemini (gemini-2.5-flash / gemini-2.5-pro)', value: 'gemini' },
|
|
10
10
|
];
|
|
11
11
|
|