@mariozechner/pi-ai 0.19.0 → 0.19.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -5745,7 +5745,7 @@ export declare const MODELS: {
5745
5745
  contextWindow: number;
5746
5746
  maxTokens: number;
5747
5747
  };
5748
- readonly "cohere/command-r-plus-08-2024": {
5748
+ readonly "cohere/command-r-08-2024": {
5749
5749
  id: string;
5750
5750
  name: string;
5751
5751
  api: "openai-completions";
@@ -5762,7 +5762,7 @@ export declare const MODELS: {
5762
5762
  contextWindow: number;
5763
5763
  maxTokens: number;
5764
5764
  };
5765
- readonly "cohere/command-r-08-2024": {
5765
+ readonly "cohere/command-r-plus-08-2024": {
5766
5766
  id: string;
5767
5767
  name: string;
5768
5768
  api: "openai-completions";
@@ -5847,7 +5847,7 @@ export declare const MODELS: {
5847
5847
  contextWindow: number;
5848
5848
  maxTokens: number;
5849
5849
  };
5850
- readonly "meta-llama/llama-3.1-70b-instruct": {
5850
+ readonly "meta-llama/llama-3.1-405b-instruct": {
5851
5851
  id: string;
5852
5852
  name: string;
5853
5853
  api: "openai-completions";
@@ -5864,7 +5864,7 @@ export declare const MODELS: {
5864
5864
  contextWindow: number;
5865
5865
  maxTokens: number;
5866
5866
  };
5867
- readonly "meta-llama/llama-3.1-405b-instruct": {
5867
+ readonly "meta-llama/llama-3.1-70b-instruct": {
5868
5868
  id: string;
5869
5869
  name: string;
5870
5870
  api: "openai-completions";
@@ -5898,7 +5898,7 @@ export declare const MODELS: {
5898
5898
  contextWindow: number;
5899
5899
  maxTokens: number;
5900
5900
  };
5901
- readonly "openai/gpt-4o-mini": {
5901
+ readonly "openai/gpt-4o-mini-2024-07-18": {
5902
5902
  id: string;
5903
5903
  name: string;
5904
5904
  api: "openai-completions";
@@ -5915,7 +5915,7 @@ export declare const MODELS: {
5915
5915
  contextWindow: number;
5916
5916
  maxTokens: number;
5917
5917
  };
5918
- readonly "openai/gpt-4o-mini-2024-07-18": {
5918
+ readonly "openai/gpt-4o-mini": {
5919
5919
  id: string;
5920
5920
  name: string;
5921
5921
  api: "openai-completions";
@@ -6289,7 +6289,7 @@ export declare const MODELS: {
6289
6289
  contextWindow: number;
6290
6290
  maxTokens: number;
6291
6291
  };
6292
- readonly "openai/gpt-4": {
6292
+ readonly "openai/gpt-4-0314": {
6293
6293
  id: string;
6294
6294
  name: string;
6295
6295
  api: "openai-completions";
@@ -6306,7 +6306,7 @@ export declare const MODELS: {
6306
6306
  contextWindow: number;
6307
6307
  maxTokens: number;
6308
6308
  };
6309
- readonly "openai/gpt-3.5-turbo": {
6309
+ readonly "openai/gpt-4": {
6310
6310
  id: string;
6311
6311
  name: string;
6312
6312
  api: "openai-completions";
@@ -6323,7 +6323,7 @@ export declare const MODELS: {
6323
6323
  contextWindow: number;
6324
6324
  maxTokens: number;
6325
6325
  };
6326
- readonly "openai/gpt-4-0314": {
6326
+ readonly "openai/gpt-3.5-turbo": {
6327
6327
  id: string;
6328
6328
  name: string;
6329
6329
  api: "openai-completions";