@mariozechner/pi-ai 0.27.1 → 0.27.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3004,6 +3004,57 @@ export declare const MODELS: {
3004
3004
  };
3005
3005
  };
3006
3006
  readonly openrouter: {
3007
+ readonly "bytedance-seed/seed-1.6-flash": {
3008
+ id: string;
3009
+ name: string;
3010
+ api: "openai-completions";
3011
+ provider: string;
3012
+ baseUrl: string;
3013
+ reasoning: true;
3014
+ input: ("image" | "text")[];
3015
+ cost: {
3016
+ input: number;
3017
+ output: number;
3018
+ cacheRead: number;
3019
+ cacheWrite: number;
3020
+ };
3021
+ contextWindow: number;
3022
+ maxTokens: number;
3023
+ };
3024
+ readonly "bytedance-seed/seed-1.6": {
3025
+ id: string;
3026
+ name: string;
3027
+ api: "openai-completions";
3028
+ provider: string;
3029
+ baseUrl: string;
3030
+ reasoning: true;
3031
+ input: ("image" | "text")[];
3032
+ cost: {
3033
+ input: number;
3034
+ output: number;
3035
+ cacheRead: number;
3036
+ cacheWrite: number;
3037
+ };
3038
+ contextWindow: number;
3039
+ maxTokens: number;
3040
+ };
3041
+ readonly "minimax/minimax-m2.1": {
3042
+ id: string;
3043
+ name: string;
3044
+ api: "openai-completions";
3045
+ provider: string;
3046
+ baseUrl: string;
3047
+ reasoning: true;
3048
+ input: "text"[];
3049
+ cost: {
3050
+ input: number;
3051
+ output: number;
3052
+ cacheRead: number;
3053
+ cacheWrite: number;
3054
+ };
3055
+ contextWindow: number;
3056
+ maxTokens: number;
3057
+ };
3007
3058
  readonly "z-ai/glm-4.7": {
3008
3059
  id: string;
3009
3060
  name: string;
@@ -6183,7 +6234,7 @@ export declare const MODELS: {
6183
6234
  contextWindow: number;
6184
6235
  maxTokens: number;
6185
6236
  };
6186
- readonly "anthropic/claude-3.5-haiku": {
6237
+ readonly "anthropic/claude-3.5-haiku-20241022": {
6187
6238
  id: string;
6188
6239
  name: string;
6189
6240
  api: "openai-completions";
@@ -6200,7 +6251,7 @@ export declare const MODELS: {
6200
6251
  contextWindow: number;
6201
6252
  maxTokens: number;
6202
6253
  };
6203
- readonly "anthropic/claude-3.5-haiku-20241022": {
6254
+ readonly "anthropic/claude-3.5-haiku": {
6204
6255
  id: string;
6205
6256
  name: string;
6206
6257
  api: "openai-completions";
@@ -6234,7 +6285,7 @@ export declare const MODELS: {
6234
6285
  contextWindow: number;
6235
6286
  maxTokens: number;
6236
6287
  };
6237
- readonly "mistralai/ministral-3b": {
6288
+ readonly "mistralai/ministral-8b": {
6238
6289
  id: string;
6239
6290
  name: string;
6240
6291
  api: "openai-completions";
@@ -6251,7 +6302,7 @@ export declare const MODELS: {
6251
6302
  contextWindow: number;
6252
6303
  maxTokens: number;
6253
6304
  };
6254
- readonly "mistralai/ministral-8b": {
6305
+ readonly "mistralai/ministral-3b": {
6255
6306
  id: string;
6256
6307
  name: string;
6257
6308
  api: "openai-completions";
@@ -6455,7 +6506,7 @@ export declare const MODELS: {
6455
6506
  contextWindow: number;
6456
6507
  maxTokens: number;
6457
6508
  };
6458
- readonly "meta-llama/llama-3.1-405b-instruct": {
6509
+ readonly "meta-llama/llama-3.1-8b-instruct": {
6459
6510
  id: string;
6460
6511
  name: string;
6461
6512
  api: "openai-completions";
@@ -6472,7 +6523,7 @@ export declare const MODELS: {
6472
6523
  contextWindow: number;
6473
6524
  maxTokens: number;
6474
6525
  };
6475
- readonly "meta-llama/llama-3.1-8b-instruct": {
6526
+ readonly "meta-llama/llama-3.1-405b-instruct": {
6476
6527
  id: string;
6477
6528
  name: string;
6478
6529
  api: "openai-completions";
@@ -6506,7 +6557,7 @@ export declare const MODELS: {
6506
6557
  contextWindow: number;
6507
6558
  maxTokens: number;
6508
6559
  };
6509
- readonly "openai/gpt-4o-mini": {
6560
+ readonly "openai/gpt-4o-mini-2024-07-18": {
6510
6561
  id: string;
6511
6562
  name: string;
6512
6563
  api: "openai-completions";
@@ -6523,7 +6574,7 @@ export declare const MODELS: {
6523
6574
  contextWindow: number;
6524
6575
  maxTokens: number;
6525
6576
  };
6526
- readonly "openai/gpt-4o-mini-2024-07-18": {
6577
+ readonly "openai/gpt-4o-mini": {
6527
6578
  id: string;
6528
6579
  name: string;
6529
6580
  api: "openai-completions";
@@ -6676,7 +6727,7 @@ export declare const MODELS: {
6676
6727
  contextWindow: number;
6677
6728
  maxTokens: number;
6678
6729
  };
6679
- readonly "meta-llama/llama-3-8b-instruct": {
6730
+ readonly "meta-llama/llama-3-70b-instruct": {
6680
6731
  id: string;
6681
6732
  name: string;
6682
6733
  api: "openai-completions";
@@ -6693,7 +6744,7 @@ export declare const MODELS: {
6693
6744
  contextWindow: number;
6694
6745
  maxTokens: number;
6695
6746
  };
6696
- readonly "meta-llama/llama-3-70b-instruct": {
6747
+ readonly "meta-llama/llama-3-8b-instruct": {
6697
6748
  id: string;
6698
6749
  name: string;
6699
6750
  api: "openai-completions";
@@ -6914,7 +6965,7 @@ export declare const MODELS: {
6914
6965
  contextWindow: number;
6915
6966
  maxTokens: number;
6916
6967
  };
6917
- readonly "openai/gpt-3.5-turbo": {
6968
+ readonly "openai/gpt-4": {
6918
6969
  id: string;
6919
6970
  name: string;
6920
6971
  api: "openai-completions";
@@ -6931,7 +6982,7 @@ export declare const MODELS: {
6931
6982
  contextWindow: number;
6932
6983
  maxTokens: number;
6933
6984
  };
6934
- readonly "openai/gpt-4": {
6985
+ readonly "openai/gpt-3.5-turbo": {
6935
6986
  id: string;
6936
6987
  name: string;
6937
6988
  api: "openai-completions";