@mariozechner/pi-ai 0.19.1 → 0.19.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3603,23 +3603,6 @@ export declare const MODELS: {
3603
3603
  contextWindow: number;
3604
3604
  maxTokens: number;
3605
3605
  };
3606
- readonly "meituan/longcat-flash-chat:free": {
3607
- id: string;
3608
- name: string;
3609
- api: "openai-completions";
3610
- provider: string;
3611
- baseUrl: string;
3612
- reasoning: false;
3613
- input: "text"[];
3614
- cost: {
3615
- input: number;
3616
- output: number;
3617
- cacheRead: number;
3618
- cacheWrite: number;
3619
- };
3620
- contextWindow: number;
3621
- maxTokens: number;
3622
- };
3623
3606
  readonly "qwen/qwen-plus-2025-07-28": {
3624
3607
  id: string;
3625
3608
  name: string;
@@ -5745,7 +5728,7 @@ export declare const MODELS: {
5745
5728
  contextWindow: number;
5746
5729
  maxTokens: number;
5747
5730
  };
5748
- readonly "cohere/command-r-08-2024": {
5731
+ readonly "cohere/command-r-plus-08-2024": {
5749
5732
  id: string;
5750
5733
  name: string;
5751
5734
  api: "openai-completions";
@@ -5762,7 +5745,7 @@ export declare const MODELS: {
5762
5745
  contextWindow: number;
5763
5746
  maxTokens: number;
5764
5747
  };
5765
- readonly "cohere/command-r-plus-08-2024": {
5748
+ readonly "cohere/command-r-08-2024": {
5766
5749
  id: string;
5767
5750
  name: string;
5768
5751
  api: "openai-completions";
@@ -5847,7 +5830,7 @@ export declare const MODELS: {
5847
5830
  contextWindow: number;
5848
5831
  maxTokens: number;
5849
5832
  };
5850
- readonly "meta-llama/llama-3.1-405b-instruct": {
5833
+ readonly "meta-llama/llama-3.1-70b-instruct": {
5851
5834
  id: string;
5852
5835
  name: string;
5853
5836
  api: "openai-completions";
@@ -5864,7 +5847,7 @@ export declare const MODELS: {
5864
5847
  contextWindow: number;
5865
5848
  maxTokens: number;
5866
5849
  };
5867
- readonly "meta-llama/llama-3.1-70b-instruct": {
5850
+ readonly "meta-llama/llama-3.1-405b-instruct": {
5868
5851
  id: string;
5869
5852
  name: string;
5870
5853
  api: "openai-completions";
@@ -5898,7 +5881,7 @@ export declare const MODELS: {
5898
5881
  contextWindow: number;
5899
5882
  maxTokens: number;
5900
5883
  };
5901
- readonly "openai/gpt-4o-mini-2024-07-18": {
5884
+ readonly "openai/gpt-4o-mini": {
5902
5885
  id: string;
5903
5886
  name: string;
5904
5887
  api: "openai-completions";
@@ -5915,7 +5898,7 @@ export declare const MODELS: {
5915
5898
  contextWindow: number;
5916
5899
  maxTokens: number;
5917
5900
  };
5918
- readonly "openai/gpt-4o-mini": {
5901
+ readonly "openai/gpt-4o-mini-2024-07-18": {
5919
5902
  id: string;
5920
5903
  name: string;
5921
5904
  api: "openai-completions";
@@ -6068,7 +6051,7 @@ export declare const MODELS: {
6068
6051
  contextWindow: number;
6069
6052
  maxTokens: number;
6070
6053
  };
6071
- readonly "meta-llama/llama-3-70b-instruct": {
6054
+ readonly "meta-llama/llama-3-8b-instruct": {
6072
6055
  id: string;
6073
6056
  name: string;
6074
6057
  api: "openai-completions";
@@ -6085,7 +6068,7 @@ export declare const MODELS: {
6085
6068
  contextWindow: number;
6086
6069
  maxTokens: number;
6087
6070
  };
6088
- readonly "meta-llama/llama-3-8b-instruct": {
6071
+ readonly "meta-llama/llama-3-70b-instruct": {
6089
6072
  id: string;
6090
6073
  name: string;
6091
6074
  api: "openai-completions";
@@ -6187,7 +6170,7 @@ export declare const MODELS: {
6187
6170
  contextWindow: number;
6188
6171
  maxTokens: number;
6189
6172
  };
6190
- readonly "openai/gpt-3.5-turbo-0613": {
6173
+ readonly "openai/gpt-4-turbo-preview": {
6191
6174
  id: string;
6192
6175
  name: string;
6193
6176
  api: "openai-completions";
@@ -6204,7 +6187,7 @@ export declare const MODELS: {
6204
6187
  contextWindow: number;
6205
6188
  maxTokens: number;
6206
6189
  };
6207
- readonly "openai/gpt-4-turbo-preview": {
6190
+ readonly "openai/gpt-3.5-turbo-0613": {
6208
6191
  id: string;
6209
6192
  name: string;
6210
6193
  api: "openai-completions";
@@ -6289,7 +6272,7 @@ export declare const MODELS: {
6289
6272
  contextWindow: number;
6290
6273
  maxTokens: number;
6291
6274
  };
6292
- readonly "openai/gpt-4-0314": {
6275
+ readonly "openai/gpt-4": {
6293
6276
  id: string;
6294
6277
  name: string;
6295
6278
  api: "openai-completions";
@@ -6306,7 +6289,7 @@ export declare const MODELS: {
6306
6289
  contextWindow: number;
6307
6290
  maxTokens: number;
6308
6291
  };
6309
- readonly "openai/gpt-4": {
6292
+ readonly "openai/gpt-4-0314": {
6310
6293
  id: string;
6311
6294
  name: string;
6312
6295
  api: "openai-completions";