@mariozechner/pi-ai 0.54.2 → 0.55.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -6046,14 +6046,14 @@ export declare const MODELS: {
6046
6046
  contextWindow: number;
6047
6047
  maxTokens: number;
6048
6048
  };
6049
- readonly "glm-5-free": {
6049
+ readonly "gpt-5": {
6050
6050
  id: string;
6051
6051
  name: string;
6052
- api: "openai-completions";
6052
+ api: "openai-responses";
6053
6053
  provider: string;
6054
6054
  baseUrl: string;
6055
6055
  reasoning: true;
6056
- input: "text"[];
6056
+ input: ("image" | "text")[];
6057
6057
  cost: {
6058
6058
  input: number;
6059
6059
  output: number;
@@ -6063,7 +6063,7 @@ export declare const MODELS: {
6063
6063
  contextWindow: number;
6064
6064
  maxTokens: number;
6065
6065
  };
6066
- readonly "gpt-5": {
6066
+ readonly "gpt-5-codex": {
6067
6067
  id: string;
6068
6068
  name: string;
6069
6069
  api: "openai-responses";
@@ -6080,7 +6080,7 @@ export declare const MODELS: {
6080
6080
  contextWindow: number;
6081
6081
  maxTokens: number;
6082
6082
  };
6083
- readonly "gpt-5-codex": {
6083
+ readonly "gpt-5-nano": {
6084
6084
  id: string;
6085
6085
  name: string;
6086
6086
  api: "openai-responses";
@@ -6097,7 +6097,7 @@ export declare const MODELS: {
6097
6097
  contextWindow: number;
6098
6098
  maxTokens: number;
6099
6099
  };
6100
- readonly "gpt-5-nano": {
6100
+ readonly "gpt-5.1": {
6101
6101
  id: string;
6102
6102
  name: string;
6103
6103
  api: "openai-responses";
@@ -6114,7 +6114,7 @@ export declare const MODELS: {
6114
6114
  contextWindow: number;
6115
6115
  maxTokens: number;
6116
6116
  };
6117
- readonly "gpt-5.1": {
6117
+ readonly "gpt-5.1-codex": {
6118
6118
  id: string;
6119
6119
  name: string;
6120
6120
  api: "openai-responses";
@@ -6131,7 +6131,7 @@ export declare const MODELS: {
6131
6131
  contextWindow: number;
6132
6132
  maxTokens: number;
6133
6133
  };
6134
- readonly "gpt-5.1-codex": {
6134
+ readonly "gpt-5.1-codex-max": {
6135
6135
  id: string;
6136
6136
  name: string;
6137
6137
  api: "openai-responses";
@@ -6148,7 +6148,7 @@ export declare const MODELS: {
6148
6148
  contextWindow: number;
6149
6149
  maxTokens: number;
6150
6150
  };
6151
- readonly "gpt-5.1-codex-max": {
6151
+ readonly "gpt-5.1-codex-mini": {
6152
6152
  id: string;
6153
6153
  name: string;
6154
6154
  api: "openai-responses";
@@ -6165,7 +6165,7 @@ export declare const MODELS: {
6165
6165
  contextWindow: number;
6166
6166
  maxTokens: number;
6167
6167
  };
6168
- readonly "gpt-5.1-codex-mini": {
6168
+ readonly "gpt-5.2": {
6169
6169
  id: string;
6170
6170
  name: string;
6171
6171
  api: "openai-responses";
@@ -6182,7 +6182,7 @@ export declare const MODELS: {
6182
6182
  contextWindow: number;
6183
6183
  maxTokens: number;
6184
6184
  };
6185
- readonly "gpt-5.2": {
6185
+ readonly "gpt-5.2-codex": {
6186
6186
  id: string;
6187
6187
  name: string;
6188
6188
  api: "openai-responses";
@@ -6199,7 +6199,7 @@ export declare const MODELS: {
6199
6199
  contextWindow: number;
6200
6200
  maxTokens: number;
6201
6201
  };
6202
- readonly "gpt-5.2-codex": {
6202
+ readonly "gpt-5.3-codex": {
6203
6203
  id: string;
6204
6204
  name: string;
6205
6205
  api: "openai-responses";
@@ -8717,6 +8717,23 @@ export declare const MODELS: {
8717
8717
  contextWindow: number;
8718
8718
  maxTokens: number;
8719
8719
  };
8720
+ readonly "openai/gpt-5.3-codex": {
8721
+ id: string;
8722
+ name: string;
8723
+ api: "openai-completions";
8724
+ provider: string;
8725
+ baseUrl: string;
8726
+ reasoning: true;
8727
+ input: ("image" | "text")[];
8728
+ cost: {
8729
+ input: number;
8730
+ output: number;
8731
+ cacheRead: number;
8732
+ cacheWrite: number;
8733
+ };
8734
+ contextWindow: number;
8735
+ maxTokens: number;
8736
+ };
8720
8737
  readonly "openai/gpt-oss-120b": {
8721
8738
  id: string;
8722
8739
  name: string;
@@ -10402,6 +10419,23 @@ export declare const MODELS: {
10402
10419
  contextWindow: number;
10403
10420
  maxTokens: number;
10404
10421
  };
10422
+ readonly "alibaba/qwen3.5-flash": {
10423
+ id: string;
10424
+ name: string;
10425
+ api: "anthropic-messages";
10426
+ provider: string;
10427
+ baseUrl: string;
10428
+ reasoning: true;
10429
+ input: ("image" | "text")[];
10430
+ cost: {
10431
+ input: number;
10432
+ output: number;
10433
+ cacheRead: number;
10434
+ cacheWrite: number;
10435
+ };
10436
+ contextWindow: number;
10437
+ maxTokens: number;
10438
+ };
10405
10439
  readonly "alibaba/qwen3.5-plus": {
10406
10440
  id: string;
10407
10441
  name: string;
@@ -11813,6 +11847,23 @@ export declare const MODELS: {
11813
11847
  contextWindow: number;
11814
11848
  maxTokens: number;
11815
11849
  };
11850
+ readonly "openai/gpt-5.3-codex": {
11851
+ id: string;
11852
+ name: string;
11853
+ api: "anthropic-messages";
11854
+ provider: string;
11855
+ baseUrl: string;
11856
+ reasoning: true;
11857
+ input: ("image" | "text")[];
11858
+ cost: {
11859
+ input: number;
11860
+ output: number;
11861
+ cacheRead: number;
11862
+ cacheWrite: number;
11863
+ };
11864
+ contextWindow: number;
11865
+ maxTokens: number;
11866
+ };
11816
11867
  readonly "openai/gpt-oss-120b": {
11817
11868
  id: string;
11818
11869
  name: string;