@oh-my-pi/pi-ai 8.2.2 → 8.3.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +2 -2
- package/src/models.generated.ts +4 -38
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@oh-my-pi/pi-ai",
|
|
3
|
-
"version": "8.
|
|
3
|
+
"version": "8.3.0",
|
|
4
4
|
"description": "Unified LLM API with automatic model discovery and provider configuration",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"main": "./src/index.ts",
|
|
@@ -56,7 +56,7 @@
|
|
|
56
56
|
"test": "bun test"
|
|
57
57
|
},
|
|
58
58
|
"dependencies": {
|
|
59
|
-
"@oh-my-pi/pi-utils": "8.
|
|
59
|
+
"@oh-my-pi/pi-utils": "8.3.0",
|
|
60
60
|
"@anthropic-ai/sdk": "^0.71.2",
|
|
61
61
|
"@aws-sdk/client-bedrock-runtime": "^3.975.0",
|
|
62
62
|
"@bufbuild/protobuf": "^2.10.2",
|
package/src/models.generated.ts
CHANGED
|
@@ -4092,40 +4092,6 @@ export const MODELS = {
|
|
|
4092
4092
|
} satisfies Model<"openai-codex-responses">,
|
|
4093
4093
|
},
|
|
4094
4094
|
"opencode": {
|
|
4095
|
-
"alpha-gd4": {
|
|
4096
|
-
id: "alpha-gd4",
|
|
4097
|
-
name: "Alpha GD4",
|
|
4098
|
-
api: "anthropic-messages",
|
|
4099
|
-
provider: "opencode",
|
|
4100
|
-
baseUrl: "https://opencode.ai/zen",
|
|
4101
|
-
reasoning: true,
|
|
4102
|
-
input: ["text"],
|
|
4103
|
-
cost: {
|
|
4104
|
-
input: 0.5,
|
|
4105
|
-
output: 2,
|
|
4106
|
-
cacheRead: 0.15,
|
|
4107
|
-
cacheWrite: 0,
|
|
4108
|
-
},
|
|
4109
|
-
contextWindow: 262144,
|
|
4110
|
-
maxTokens: 32768,
|
|
4111
|
-
} satisfies Model<"anthropic-messages">,
|
|
4112
|
-
"alpha-glm-4.7": {
|
|
4113
|
-
id: "alpha-glm-4.7",
|
|
4114
|
-
name: "Alpha GLM-4.7",
|
|
4115
|
-
api: "openai-completions",
|
|
4116
|
-
provider: "opencode",
|
|
4117
|
-
baseUrl: "https://opencode.ai/zen/v1",
|
|
4118
|
-
reasoning: true,
|
|
4119
|
-
input: ["text"],
|
|
4120
|
-
cost: {
|
|
4121
|
-
input: 0.6,
|
|
4122
|
-
output: 2.2,
|
|
4123
|
-
cacheRead: 0.6,
|
|
4124
|
-
cacheWrite: 0,
|
|
4125
|
-
},
|
|
4126
|
-
contextWindow: 204800,
|
|
4127
|
-
maxTokens: 131072,
|
|
4128
|
-
} satisfies Model<"openai-completions">,
|
|
4129
4095
|
"big-pickle": {
|
|
4130
4096
|
id: "big-pickle",
|
|
4131
4097
|
name: "Big Pickle",
|
|
@@ -5310,10 +5276,10 @@ export const MODELS = {
|
|
|
5310
5276
|
reasoning: false,
|
|
5311
5277
|
input: ["text", "image"],
|
|
5312
5278
|
cost: {
|
|
5313
|
-
input: 0
|
|
5314
|
-
output: 0
|
|
5315
|
-
cacheRead: 0
|
|
5316
|
-
cacheWrite: 0
|
|
5279
|
+
input: 0,
|
|
5280
|
+
output: 0,
|
|
5281
|
+
cacheRead: 0,
|
|
5282
|
+
cacheWrite: 0,
|
|
5317
5283
|
},
|
|
5318
5284
|
contextWindow: 1048576,
|
|
5319
5285
|
maxTokens: 8192,
|