@oh-my-pi/pi-ai 8.2.2 → 8.3.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@oh-my-pi/pi-ai",
3
- "version": "8.2.2",
3
+ "version": "8.3.0",
4
4
  "description": "Unified LLM API with automatic model discovery and provider configuration",
5
5
  "type": "module",
6
6
  "main": "./src/index.ts",
@@ -56,7 +56,7 @@
56
56
  "test": "bun test"
57
57
  },
58
58
  "dependencies": {
59
- "@oh-my-pi/pi-utils": "8.2.2",
59
+ "@oh-my-pi/pi-utils": "8.3.0",
60
60
  "@anthropic-ai/sdk": "^0.71.2",
61
61
  "@aws-sdk/client-bedrock-runtime": "^3.975.0",
62
62
  "@bufbuild/protobuf": "^2.10.2",
@@ -4092,40 +4092,6 @@ export const MODELS = {
4092
4092
  } satisfies Model<"openai-codex-responses">,
4093
4093
  },
4094
4094
  "opencode": {
4095
- "alpha-gd4": {
4096
- id: "alpha-gd4",
4097
- name: "Alpha GD4",
4098
- api: "anthropic-messages",
4099
- provider: "opencode",
4100
- baseUrl: "https://opencode.ai/zen",
4101
- reasoning: true,
4102
- input: ["text"],
4103
- cost: {
4104
- input: 0.5,
4105
- output: 2,
4106
- cacheRead: 0.15,
4107
- cacheWrite: 0,
4108
- },
4109
- contextWindow: 262144,
4110
- maxTokens: 32768,
4111
- } satisfies Model<"anthropic-messages">,
4112
- "alpha-glm-4.7": {
4113
- id: "alpha-glm-4.7",
4114
- name: "Alpha GLM-4.7",
4115
- api: "openai-completions",
4116
- provider: "opencode",
4117
- baseUrl: "https://opencode.ai/zen/v1",
4118
- reasoning: true,
4119
- input: ["text"],
4120
- cost: {
4121
- input: 0.6,
4122
- output: 2.2,
4123
- cacheRead: 0.6,
4124
- cacheWrite: 0,
4125
- },
4126
- contextWindow: 204800,
4127
- maxTokens: 131072,
4128
- } satisfies Model<"openai-completions">,
4129
4095
  "big-pickle": {
4130
4096
  id: "big-pickle",
4131
4097
  name: "Big Pickle",
@@ -5310,10 +5276,10 @@ export const MODELS = {
5310
5276
  reasoning: false,
5311
5277
  input: ["text", "image"],
5312
5278
  cost: {
5313
- input: 0.09999999999999999,
5314
- output: 0.39999999999999997,
5315
- cacheRead: 0.024999999999999998,
5316
- cacheWrite: 0.08333333333333334,
5279
+ input: 0,
5280
+ output: 0,
5281
+ cacheRead: 0,
5282
+ cacheWrite: 0,
5317
5283
  },
5318
5284
  contextWindow: 1048576,
5319
5285
  maxTokens: 8192,