@mariozechner/pi-ai 0.9.2 → 0.9.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -102,6 +102,23 @@ export declare const MODELS: {
102
102
  contextWindow: number;
103
103
  maxTokens: number;
104
104
  };
105
+ readonly "claude-opus-4-5": {
106
+ id: string;
107
+ name: string;
108
+ api: "anthropic-messages";
109
+ provider: string;
110
+ baseUrl: string;
111
+ reasoning: true;
112
+ input: ("image" | "text")[];
113
+ cost: {
114
+ input: number;
115
+ output: number;
116
+ cacheRead: number;
117
+ cacheWrite: number;
118
+ };
119
+ contextWindow: number;
120
+ maxTokens: number;
121
+ };
105
122
  readonly "claude-3-opus-20240229": {
106
123
  id: string;
107
124
  name: string;
@@ -1935,14 +1952,14 @@ export declare const MODELS: {
1935
1952
  };
1936
1953
  };
1937
1954
  readonly openrouter: {
1938
- readonly "allenai/olmo-3-7b-instruct": {
1955
+ readonly "anthropic/claude-opus-4.5": {
1939
1956
  id: string;
1940
1957
  name: string;
1941
1958
  api: "openai-completions";
1942
1959
  provider: string;
1943
1960
  baseUrl: string;
1944
- reasoning: false;
1945
- input: "text"[];
1961
+ reasoning: true;
1962
+ input: ("image" | "text")[];
1946
1963
  cost: {
1947
1964
  input: number;
1948
1965
  output: number;
@@ -1952,14 +1969,14 @@ export declare const MODELS: {
1952
1969
  contextWindow: number;
1953
1970
  maxTokens: number;
1954
1971
  };
1955
- readonly "x-ai/grok-4.1-fast": {
1972
+ readonly "allenai/olmo-3-7b-instruct": {
1956
1973
  id: string;
1957
1974
  name: string;
1958
1975
  api: "openai-completions";
1959
1976
  provider: string;
1960
1977
  baseUrl: string;
1961
- reasoning: true;
1962
- input: ("image" | "text")[];
1978
+ reasoning: false;
1979
+ input: "text"[];
1963
1980
  cost: {
1964
1981
  input: number;
1965
1982
  output: number;
@@ -4791,7 +4808,7 @@ export declare const MODELS: {
4791
4808
  contextWindow: number;
4792
4809
  maxTokens: number;
4793
4810
  };
4794
- readonly "anthropic/claude-3.5-haiku-20241022": {
4811
+ readonly "anthropic/claude-3.5-haiku": {
4795
4812
  id: string;
4796
4813
  name: string;
4797
4814
  api: "openai-completions";
@@ -4808,7 +4825,7 @@ export declare const MODELS: {
4808
4825
  contextWindow: number;
4809
4826
  maxTokens: number;
4810
4827
  };
4811
- readonly "anthropic/claude-3.5-haiku": {
4828
+ readonly "anthropic/claude-3.5-haiku-20241022": {
4812
4829
  id: string;
4813
4830
  name: string;
4814
4831
  api: "openai-completions";
@@ -4876,23 +4893,6 @@ export declare const MODELS: {
4876
4893
  contextWindow: number;
4877
4894
  maxTokens: number;
4878
4895
  };
4879
- readonly "qwen/qwen-2.5-7b-instruct": {
4880
- id: string;
4881
- name: string;
4882
- api: "openai-completions";
4883
- provider: string;
4884
- baseUrl: string;
4885
- reasoning: false;
4886
- input: "text"[];
4887
- cost: {
4888
- input: number;
4889
- output: number;
4890
- cacheRead: number;
4891
- cacheWrite: number;
4892
- };
4893
- contextWindow: number;
4894
- maxTokens: number;
4895
- };
4896
4896
  readonly "nvidia/llama-3.1-nemotron-70b-instruct": {
4897
4897
  id: string;
4898
4898
  name: string;
@@ -5080,7 +5080,7 @@ export declare const MODELS: {
5080
5080
  contextWindow: number;
5081
5081
  maxTokens: number;
5082
5082
  };
5083
- readonly "meta-llama/llama-3.1-70b-instruct": {
5083
+ readonly "meta-llama/llama-3.1-8b-instruct": {
5084
5084
  id: string;
5085
5085
  name: string;
5086
5086
  api: "openai-completions";
@@ -5097,7 +5097,7 @@ export declare const MODELS: {
5097
5097
  contextWindow: number;
5098
5098
  maxTokens: number;
5099
5099
  };
5100
- readonly "meta-llama/llama-3.1-8b-instruct": {
5100
+ readonly "meta-llama/llama-3.1-405b-instruct": {
5101
5101
  id: string;
5102
5102
  name: string;
5103
5103
  api: "openai-completions";
@@ -5114,7 +5114,7 @@ export declare const MODELS: {
5114
5114
  contextWindow: number;
5115
5115
  maxTokens: number;
5116
5116
  };
5117
- readonly "meta-llama/llama-3.1-405b-instruct": {
5117
+ readonly "meta-llama/llama-3.1-70b-instruct": {
5118
5118
  id: string;
5119
5119
  name: string;
5120
5120
  api: "openai-completions";
@@ -5148,7 +5148,7 @@ export declare const MODELS: {
5148
5148
  contextWindow: number;
5149
5149
  maxTokens: number;
5150
5150
  };
5151
- readonly "openai/gpt-4o-mini-2024-07-18": {
5151
+ readonly "openai/gpt-4o-mini": {
5152
5152
  id: string;
5153
5153
  name: string;
5154
5154
  api: "openai-completions";
@@ -5165,7 +5165,7 @@ export declare const MODELS: {
5165
5165
  contextWindow: number;
5166
5166
  maxTokens: number;
5167
5167
  };
5168
- readonly "openai/gpt-4o-mini": {
5168
+ readonly "openai/gpt-4o-mini-2024-07-18": {
5169
5169
  id: string;
5170
5170
  name: string;
5171
5171
  api: "openai-completions";
@@ -5267,7 +5267,7 @@ export declare const MODELS: {
5267
5267
  contextWindow: number;
5268
5268
  maxTokens: number;
5269
5269
  };
5270
- readonly "openai/gpt-4o-2024-05-13": {
5270
+ readonly "openai/gpt-4o": {
5271
5271
  id: string;
5272
5272
  name: string;
5273
5273
  api: "openai-completions";
@@ -5284,7 +5284,7 @@ export declare const MODELS: {
5284
5284
  contextWindow: number;
5285
5285
  maxTokens: number;
5286
5286
  };
5287
- readonly "openai/gpt-4o": {
5287
+ readonly "openai/gpt-4o:extended": {
5288
5288
  id: string;
5289
5289
  name: string;
5290
5290
  api: "openai-completions";
@@ -5301,7 +5301,7 @@ export declare const MODELS: {
5301
5301
  contextWindow: number;
5302
5302
  maxTokens: number;
5303
5303
  };
5304
- readonly "openai/gpt-4o:extended": {
5304
+ readonly "openai/gpt-4o-2024-05-13": {
5305
5305
  id: string;
5306
5306
  name: string;
5307
5307
  api: "openai-completions";
@@ -5318,7 +5318,7 @@ export declare const MODELS: {
5318
5318
  contextWindow: number;
5319
5319
  maxTokens: number;
5320
5320
  };
5321
- readonly "meta-llama/llama-3-70b-instruct": {
5321
+ readonly "meta-llama/llama-3-8b-instruct": {
5322
5322
  id: string;
5323
5323
  name: string;
5324
5324
  api: "openai-completions";
@@ -5335,7 +5335,7 @@ export declare const MODELS: {
5335
5335
  contextWindow: number;
5336
5336
  maxTokens: number;
5337
5337
  };
5338
- readonly "meta-llama/llama-3-8b-instruct": {
5338
+ readonly "meta-llama/llama-3-70b-instruct": {
5339
5339
  id: string;
5340
5340
  name: string;
5341
5341
  api: "openai-completions";