opencode-aicodewith-auth 0.1.38 → 0.1.39

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (3) hide show
  1. package/README.md +1 -1
  2. package/dist/index.js +28 -17
  3. package/package.json +2 -2
package/README.md CHANGED
@@ -61,7 +61,7 @@ User Request → OpenCode → Plugin Auth Hook → Route by Model:
61
61
 
62
62
  | 模型 ID | 显示名称 | 图片输入 | 适合场景 |
63
63
  |---------|---------|:-------:|---------|
64
- | `aicodewith/gpt-5.2-codex` | GPT-5.2 Codex | ✅ | 日常编程、代码生成 |
64
+ | `aicodewith/gpt-5.3-codex` | GPT-5.3 Codex | ✅ | 日常编程、代码生成 |
65
65
  | `aicodewith/gpt-5.2` | GPT-5.2 | ✅ | 架构设计、逻辑推理 |
66
66
  | `aicodewith/claude-sonnet-4-5-20250929` | Claude Sonnet 4.5 | ✅ | 代码审查、文档查询 |
67
67
  | `aicodewith/claude-opus-4-6-20260205` | Claude Opus 4.6 | ✅ | 复杂任务、深度思考 |
package/dist/index.js CHANGED
@@ -8,14 +8,14 @@ import os4 from "os";
8
8
  var PROVIDER_ID = "aicodewith";
9
9
  var MODELS = [
10
10
  {
11
- id: "gpt-5.2-codex",
11
+ id: "gpt-5.3-codex",
12
12
  family: "codex",
13
- displayName: "GPT-5.2 Codex",
14
- version: "5.2",
13
+ displayName: "GPT-5.3 Codex",
14
+ version: "5.3",
15
15
  limit: { context: 400000, output: 128000 },
16
16
  modalities: { input: ["text", "image"], output: ["text"] },
17
17
  reasoning: "xhigh",
18
- aliases: ["gpt-5.2-codex", "gpt 5.2 codex", "codex"]
18
+ aliases: ["gpt-5.3-codex", "gpt 5.3 codex", "codex"]
19
19
  },
20
20
  {
21
21
  id: "gpt-5.2",
@@ -27,6 +27,17 @@ var MODELS = [
27
27
  reasoning: "xhigh",
28
28
  aliases: ["gpt-5.2", "gpt 5.2"]
29
29
  },
30
+ {
31
+ id: "gpt-5.2-codex",
32
+ family: "codex",
33
+ displayName: "GPT-5.2 Codex (deprecated)",
34
+ version: "5.2",
35
+ limit: { context: 400000, output: 128000 },
36
+ modalities: { input: ["text", "image"], output: ["text"] },
37
+ reasoning: "xhigh",
38
+ deprecated: true,
39
+ replacedBy: "gpt-5.3-codex"
40
+ },
30
41
  {
31
42
  id: "gpt-5.1-codex",
32
43
  family: "codex",
@@ -36,7 +47,7 @@ var MODELS = [
36
47
  modalities: { input: ["text", "image"], output: ["text"] },
37
48
  reasoning: "full",
38
49
  deprecated: true,
39
- replacedBy: "gpt-5.2-codex"
50
+ replacedBy: "gpt-5.3-codex"
40
51
  },
41
52
  {
42
53
  id: "gpt-5.1-codex-max",
@@ -47,7 +58,7 @@ var MODELS = [
47
58
  modalities: { input: ["text", "image"], output: ["text"] },
48
59
  reasoning: "xhigh",
49
60
  deprecated: true,
50
- replacedBy: "gpt-5.2-codex"
61
+ replacedBy: "gpt-5.3-codex"
51
62
  },
52
63
  {
53
64
  id: "gpt-5.1-codex-mini",
@@ -58,7 +69,7 @@ var MODELS = [
58
69
  modalities: { input: ["text", "image"], output: ["text"] },
59
70
  reasoning: "basic",
60
71
  deprecated: true,
61
- replacedBy: "gpt-5.2-codex"
72
+ replacedBy: "gpt-5.3-codex"
62
73
  },
63
74
  {
64
75
  id: "gpt-5.1",
@@ -817,23 +828,23 @@ var normalizeOrphanedToolOutputs = (input) => {
817
828
  // lib/request/request-transformer.ts
818
829
  function normalizeModel(model) {
819
830
  if (!model)
820
- return "gpt-5.2-codex";
831
+ return "gpt-5.3-codex";
821
832
  const modelId = model.includes("/") ? model.split("/").pop() : model;
822
833
  const mappedModel = getNormalizedModel(modelId);
823
834
  if (mappedModel) {
824
835
  return mappedModel;
825
836
  }
826
837
  const normalized = modelId.toLowerCase();
827
- if (normalized.includes("gpt-5.2-codex") || normalized.includes("gpt 5.2 codex")) {
828
- return "gpt-5.2-codex";
838
+ if (normalized.includes("gpt-5.3-codex") || normalized.includes("gpt 5.3 codex")) {
839
+ return "gpt-5.3-codex";
829
840
  }
830
841
  if (normalized.includes("gpt-5.2") || normalized.includes("gpt 5.2")) {
831
842
  return "gpt-5.2";
832
843
  }
833
844
  if (normalized.includes("codex")) {
834
- return "gpt-5.2-codex";
845
+ return "gpt-5.3-codex";
835
846
  }
836
- return "gpt-5.2-codex";
847
+ return "gpt-5.3-codex";
837
848
  }
838
849
  function resolveReasoningConfig(modelName, body) {
839
850
  const providerOpenAI = body.providerOptions?.openai;
@@ -900,9 +911,9 @@ function addCodexBridgeMessage(input, hasTools) {
900
911
  }
901
912
  function getReasoningConfig(modelName, userConfig = {}) {
902
913
  const normalizedName = modelName?.toLowerCase() ?? "";
903
- const isGpt52Codex = normalizedName.includes("gpt-5.2-codex") || normalizedName.includes("gpt 5.2 codex");
904
- const isGpt52General = (normalizedName.includes("gpt-5.2") || normalizedName.includes("gpt 5.2")) && !isGpt52Codex;
905
- const supportsXhigh = isGpt52General || isGpt52Codex;
914
+ const isGpt53Codex = normalizedName.includes("gpt-5.3-codex") || normalizedName.includes("gpt 5.3 codex");
915
+ const isGpt52General = normalizedName.includes("gpt-5.2") || normalizedName.includes("gpt 5.2");
916
+ const supportsXhigh = isGpt52General || isGpt53Codex;
906
917
  const supportsNone = isGpt52General;
907
918
  const defaultEffort = supportsXhigh ? "high" : "medium";
908
919
  let effort = userConfig.reasoningEffort || defaultEffort;
@@ -1789,8 +1800,8 @@ var provider_config_default = {
1789
1800
  "AICODEWITH_API_KEY"
1790
1801
  ],
1791
1802
  models: {
1792
- "gpt-5.2-codex": {
1793
- name: "GPT-5.2 Codex",
1803
+ "gpt-5.3-codex": {
1804
+ name: "GPT-5.3 Codex",
1794
1805
  limit: {
1795
1806
  context: 400000,
1796
1807
  output: 128000
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "opencode-aicodewith-auth",
3
- "version": "0.1.38",
4
- "description": "OpenCode plugin for AICodewith authentication - Access GPT-5.2, Claude, and Gemini models through AICodewith API",
3
+ "version": "0.1.39",
4
+ "description": "OpenCode plugin for AICodewith authentication - Access GPT-5.3 Codex, GPT-5.2, Claude, and Gemini models through AICodewith API",
5
5
  "type": "module",
6
6
  "main": "dist/index.js",
7
7
  "types": "dist/index.d.ts",