omnikey-cli 1.0.38 → 1.0.39

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -16,9 +16,9 @@ const config_1 = require("./config");
16
16
  // Default model mapping
17
17
  // ---------------------------------------------------------------------------
18
18
  const DEFAULT_MODELS = {
19
- openai: { fast: 'gpt-4o-mini', smart: 'gpt-5.1' },
19
+ openai: { fast: 'gpt-4o-mini', smart: 'gpt-5.5' },
20
20
  gemini: { fast: 'gemini-2.5-flash', smart: 'gemini-2.5-pro' },
21
- anthropic: { fast: 'claude-haiku-4-5-20251001', smart: 'claude-sonnet-4-6' },
21
+ anthropic: { fast: 'claude-haiku-4-5-20251001', smart: 'claude-opus-4-7' },
22
22
  };
23
23
  function getDefaultModel(provider, tier) {
24
24
  return DEFAULT_MODELS[provider][tier];
@@ -29,7 +29,7 @@ function getDefaultModel(provider, tier) {
29
29
  * - anthropic: hard API-enforced string limit of 10,485,760 chars; we stay
30
30
  * just below it with a small safety buffer.
31
31
  * - openai: no documented per-string limit; bounded by the context window
32
- * (~272K tokens for GPT-5.1 ≈ ~1M chars). Use the history cap.
32
+ * (~272K tokens for GPT-5.5 ≈ ~1M chars). Use the history cap.
33
33
  * - gemini: no documented per-string limit; bounded by the 1M-token
34
34
  * context window (~4M chars). Use the history cap.
35
35
  */
@@ -43,9 +43,9 @@ const MAX_MESSAGE_CONTENT_LENGTH_BY_PROVIDER = {
43
43
  * history, derived from each provider's context-window size minus headroom
44
44
  * for the system prompt and max output tokens.
45
45
  *
46
- * - anthropic: Claude Sonnet 4.6 — 1M token ctx, 64K max output
46
+ * - anthropic: Claude Opus 4.7 — 1M token ctx, 64K max output
47
47
  * ≈ (1,000,000 - 64,000 - 10,000) tokens × 4 chars ≈ 3.7M chars
48
- * - openai: GPT-5.1 — ~272K token ctx, ~32K max output
48
+ * - openai: GPT-5.5 — ~272K token ctx, ~32K max output
49
49
  * ≈ (272,000 - 32,000 - 5,000) tokens × 4 chars ≈ 940K chars
50
50
  * - gemini: Gemini 2.5 Pro — 1M token ctx, ~32K max output
51
51
  * ≈ (1,000,000 - 32,000 - 10,000) tokens × 4 chars ≈ 3.8M chars
@@ -75,7 +75,7 @@ class OpenAIAdapter {
75
75
  model,
76
76
  messages: oaiMessages,
77
77
  tools: tools?.length ? tools : undefined,
78
- temperature: options.temperature ?? 0.2,
78
+ temperature: model === 'gpt-5.5' ? 1 : (options.temperature ?? 0.2),
79
79
  max_tokens: options.maxTokens,
80
80
  });
81
81
  const choice = completion.choices[0];
@@ -175,7 +175,7 @@ class AnthropicAdapter {
175
175
  ...(system ? { system } : {}),
176
176
  messages: anthropicMessages,
177
177
  ...(tools?.length ? { tools } : {}),
178
- temperature: options.temperature ?? 0.2,
178
+ ...(model === 'claude-opus-4-7' ? {} : { temperature: options.temperature ?? 0.2 }),
179
179
  });
180
180
  const textContent = response.content
181
181
  .filter((b) => b.type === 'text')
@@ -58,9 +58,9 @@ async function getPromptForCommand(logger, cmd, subscription) {
58
58
  function getModelForCommand(cmd) {
59
59
  const tier = cmd === 'task' ? 'smart' : 'fast';
60
60
  const models = {
61
- openai: { fast: 'gpt-4o-mini', smart: 'gpt-5.1' },
61
+ openai: { fast: 'gpt-4o-mini', smart: 'gpt-5.5' },
62
62
  gemini: { fast: 'gemini-2.5-flash', smart: 'gemini-2.5-pro' },
63
- anthropic: { fast: 'claude-haiku-4-5-20251001', smart: 'claude-sonnet-4-6' },
63
+ anthropic: { fast: 'claude-haiku-4-5-20251001', smart: 'claude-opus-4-7' },
64
64
  };
65
65
  return models[config_1.config.aiProvider]?.[tier] ?? 'gpt-4o-mini';
66
66
  }
@@ -74,8 +74,8 @@ app.get('/macos/appcast', (req, res) => {
74
74
  const appcastUrl = `${baseUrl}/macos/appcast`;
75
75
  // These should match the values embedded into the macOS app
76
76
  // Info.plist in macOS/build_release_dmg.sh.
77
- const bundleVersion = '25';
78
- const shortVersion = '1.0.24';
77
+ const bundleVersion = '26';
78
+ const shortVersion = '1.0.25';
79
79
  const xml = `<?xml version="1.0" encoding="utf-8"?>
80
80
  <rss version="2.0"
81
81
  xmlns:sparkle="http://www.andymatuschak.org/xml-namespaces/sparkle"
package/dist/onboard.js CHANGED
@@ -9,8 +9,8 @@ const fs_1 = __importDefault(require("fs"));
9
9
  const path_1 = __importDefault(require("path"));
10
10
  const utils_1 = require("./utils");
11
11
  const AI_PROVIDERS = [
12
- { name: 'OpenAI (gpt-4o-mini / gpt-5.1)', value: 'openai' },
13
- { name: 'Anthropic — Claude (claude-haiku / claude-sonnet)', value: 'anthropic' },
12
+ { name: 'OpenAI (gpt-4o-mini / gpt-5.5)', value: 'openai' },
13
+ { name: 'Anthropic — Claude (claude-haiku / claude-opus)', value: 'anthropic' },
14
14
  { name: 'Google Gemini (gemini-2.5-flash / gemini-2.5-pro)', value: 'gemini' },
15
15
  ];
16
16
  const SEARCH_PROVIDERS = [
package/package.json CHANGED
@@ -4,7 +4,7 @@
4
4
  "access": "public",
5
5
  "registry": "https://registry.npmjs.org/"
6
6
  },
7
- "version": "1.0.38",
7
+ "version": "1.0.39",
8
8
  "description": "CLI for onboarding users to Omnikey AI and configuring OPENAI_API_KEY. Use Yarn for install/build.",
9
9
  "engines": {
10
10
  "node": ">=14.0.0",
@@ -27,7 +27,7 @@
27
27
  "author": "Gurinder Rawala",
28
28
  "license": "MIT",
29
29
  "dependencies": {
30
- "@anthropic-ai/sdk": "^0.80.0",
30
+ "@anthropic-ai/sdk": "^0.96.0",
31
31
  "@google-cloud/storage": "^7.19.0",
32
32
  "@google/genai": "^1.46.0",
33
33
  "axios": "^1.13.5",
@@ -38,7 +38,7 @@
38
38
  "express": "^4.21.2",
39
39
  "inquirer": "^9.0.0",
40
40
  "jsonwebtoken": "^9.0.3",
41
- "openai": "^6.16.0",
41
+ "openai": "^6.37.0",
42
42
  "pg": "^8.18.0",
43
43
  "pg-hstore": "^2.3.4",
44
44
  "playwright-core": "^1.50.0",
package/src/onboard.ts CHANGED
@@ -4,8 +4,8 @@ import path from 'path';
4
4
  import { getConfigDir, getConfigPath } from './utils';
5
5
 
6
6
  const AI_PROVIDERS = [
7
- { name: 'OpenAI (gpt-4o-mini / gpt-5.1)', value: 'openai' },
8
- { name: 'Anthropic — Claude (claude-haiku / claude-sonnet)', value: 'anthropic' },
7
+ { name: 'OpenAI (gpt-4o-mini / gpt-5.5)', value: 'openai' },
8
+ { name: 'Anthropic — Claude (claude-haiku / claude-opus)', value: 'anthropic' },
9
9
  { name: 'Google Gemini (gemini-2.5-flash / gemini-2.5-pro)', value: 'gemini' },
10
10
  ];
11
11