@mariozechner/pi-ai 0.7.26 → 0.7.28

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1969,6 +1969,23 @@ export declare const MODELS: {
1969
1969
  contextWindow: number;
1970
1970
  maxTokens: number;
1971
1971
  };
1972
+ readonly "x-ai/grok-4.1-fast:free": {
1973
+ id: string;
1974
+ name: string;
1975
+ api: "openai-completions";
1976
+ provider: string;
1977
+ baseUrl: string;
1978
+ reasoning: true;
1979
+ input: ("image" | "text")[];
1980
+ cost: {
1981
+ input: number;
1982
+ output: number;
1983
+ cacheRead: number;
1984
+ cacheWrite: number;
1985
+ };
1986
+ contextWindow: number;
1987
+ maxTokens: number;
1988
+ };
1972
1989
  readonly "google/gemini-3-pro-preview": {
1973
1990
  id: string;
1974
1991
  name: string;
@@ -4825,7 +4842,7 @@ export declare const MODELS: {
4825
4842
  contextWindow: number;
4826
4843
  maxTokens: number;
4827
4844
  };
4828
- readonly "mistralai/ministral-3b": {
4845
+ readonly "mistralai/ministral-8b": {
4829
4846
  id: string;
4830
4847
  name: string;
4831
4848
  api: "openai-completions";
@@ -4842,7 +4859,7 @@ export declare const MODELS: {
4842
4859
  contextWindow: number;
4843
4860
  maxTokens: number;
4844
4861
  };
4845
- readonly "mistralai/ministral-8b": {
4862
+ readonly "mistralai/ministral-3b": {
4846
4863
  id: string;
4847
4864
  name: string;
4848
4865
  api: "openai-completions";
@@ -5063,7 +5080,7 @@ export declare const MODELS: {
5063
5080
  contextWindow: number;
5064
5081
  maxTokens: number;
5065
5082
  };
5066
- readonly "meta-llama/llama-3.1-70b-instruct": {
5083
+ readonly "meta-llama/llama-3.1-8b-instruct": {
5067
5084
  id: string;
5068
5085
  name: string;
5069
5086
  api: "openai-completions";
@@ -5080,7 +5097,7 @@ export declare const MODELS: {
5080
5097
  contextWindow: number;
5081
5098
  maxTokens: number;
5082
5099
  };
5083
- readonly "meta-llama/llama-3.1-405b-instruct": {
5100
+ readonly "meta-llama/llama-3.1-70b-instruct": {
5084
5101
  id: string;
5085
5102
  name: string;
5086
5103
  api: "openai-completions";
@@ -5097,7 +5114,7 @@ export declare const MODELS: {
5097
5114
  contextWindow: number;
5098
5115
  maxTokens: number;
5099
5116
  };
5100
- readonly "meta-llama/llama-3.1-8b-instruct": {
5117
+ readonly "meta-llama/llama-3.1-405b-instruct": {
5101
5118
  id: string;
5102
5119
  name: string;
5103
5120
  api: "openai-completions";