@mariozechner/pi-ai 0.63.1 → 0.63.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/models.generated.d.ts +102 -51
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +106 -55
- package/dist/models.generated.js.map +1 -1
- package/package.json +1 -1
|
@@ -3606,6 +3606,23 @@ export declare const MODELS: {
|
|
|
3606
3606
|
contextWindow: number;
|
|
3607
3607
|
maxTokens: number;
|
|
3608
3608
|
};
|
|
3609
|
+
readonly "gemma-3-27b-it": {
|
|
3610
|
+
id: string;
|
|
3611
|
+
name: string;
|
|
3612
|
+
api: "google-generative-ai";
|
|
3613
|
+
provider: string;
|
|
3614
|
+
baseUrl: string;
|
|
3615
|
+
reasoning: false;
|
|
3616
|
+
input: ("image" | "text")[];
|
|
3617
|
+
cost: {
|
|
3618
|
+
input: number;
|
|
3619
|
+
output: number;
|
|
3620
|
+
cacheRead: number;
|
|
3621
|
+
cacheWrite: number;
|
|
3622
|
+
};
|
|
3623
|
+
contextWindow: number;
|
|
3624
|
+
maxTokens: number;
|
|
3625
|
+
};
|
|
3609
3626
|
};
|
|
3610
3627
|
readonly "google-antigravity": {
|
|
3611
3628
|
readonly "claude-opus-4-5-thinking": {
|
|
@@ -4124,6 +4141,40 @@ export declare const MODELS: {
|
|
|
4124
4141
|
contextWindow: number;
|
|
4125
4142
|
maxTokens: number;
|
|
4126
4143
|
};
|
|
4144
|
+
readonly "groq/compound": {
|
|
4145
|
+
id: string;
|
|
4146
|
+
name: string;
|
|
4147
|
+
api: "openai-completions";
|
|
4148
|
+
provider: string;
|
|
4149
|
+
baseUrl: string;
|
|
4150
|
+
reasoning: true;
|
|
4151
|
+
input: "text"[];
|
|
4152
|
+
cost: {
|
|
4153
|
+
input: number;
|
|
4154
|
+
output: number;
|
|
4155
|
+
cacheRead: number;
|
|
4156
|
+
cacheWrite: number;
|
|
4157
|
+
};
|
|
4158
|
+
contextWindow: number;
|
|
4159
|
+
maxTokens: number;
|
|
4160
|
+
};
|
|
4161
|
+
readonly "groq/compound-mini": {
|
|
4162
|
+
id: string;
|
|
4163
|
+
name: string;
|
|
4164
|
+
api: "openai-completions";
|
|
4165
|
+
provider: string;
|
|
4166
|
+
baseUrl: string;
|
|
4167
|
+
reasoning: true;
|
|
4168
|
+
input: "text"[];
|
|
4169
|
+
cost: {
|
|
4170
|
+
input: number;
|
|
4171
|
+
output: number;
|
|
4172
|
+
cacheRead: number;
|
|
4173
|
+
cacheWrite: number;
|
|
4174
|
+
};
|
|
4175
|
+
contextWindow: number;
|
|
4176
|
+
maxTokens: number;
|
|
4177
|
+
};
|
|
4127
4178
|
readonly "llama-3.1-8b-instant": {
|
|
4128
4179
|
id: string;
|
|
4129
4180
|
name: string;
|
|
@@ -4311,6 +4362,23 @@ export declare const MODELS: {
|
|
|
4311
4362
|
contextWindow: number;
|
|
4312
4363
|
maxTokens: number;
|
|
4313
4364
|
};
|
|
4365
|
+
readonly "openai/gpt-oss-safeguard-20b": {
|
|
4366
|
+
id: string;
|
|
4367
|
+
name: string;
|
|
4368
|
+
api: "openai-completions";
|
|
4369
|
+
provider: string;
|
|
4370
|
+
baseUrl: string;
|
|
4371
|
+
reasoning: true;
|
|
4372
|
+
input: "text"[];
|
|
4373
|
+
cost: {
|
|
4374
|
+
input: number;
|
|
4375
|
+
output: number;
|
|
4376
|
+
cacheRead: number;
|
|
4377
|
+
cacheWrite: number;
|
|
4378
|
+
};
|
|
4379
|
+
contextWindow: number;
|
|
4380
|
+
maxTokens: number;
|
|
4381
|
+
};
|
|
4314
4382
|
readonly "qwen-qwq-32b": {
|
|
4315
4383
|
id: string;
|
|
4316
4384
|
name: string;
|
|
@@ -7700,6 +7768,23 @@ export declare const MODELS: {
|
|
|
7700
7768
|
contextWindow: number;
|
|
7701
7769
|
maxTokens: number;
|
|
7702
7770
|
};
|
|
7771
|
+
readonly "kwaipilot/kat-coder-pro-v2": {
|
|
7772
|
+
id: string;
|
|
7773
|
+
name: string;
|
|
7774
|
+
api: "openai-completions";
|
|
7775
|
+
provider: string;
|
|
7776
|
+
baseUrl: string;
|
|
7777
|
+
reasoning: false;
|
|
7778
|
+
input: "text"[];
|
|
7779
|
+
cost: {
|
|
7780
|
+
input: number;
|
|
7781
|
+
output: number;
|
|
7782
|
+
cacheRead: number;
|
|
7783
|
+
cacheWrite: number;
|
|
7784
|
+
};
|
|
7785
|
+
contextWindow: number;
|
|
7786
|
+
maxTokens: number;
|
|
7787
|
+
};
|
|
7703
7788
|
readonly "meituan/longcat-flash-chat": {
|
|
7704
7789
|
id: string;
|
|
7705
7790
|
name: string;
|
|
@@ -8227,23 +8312,6 @@ export declare const MODELS: {
|
|
|
8227
8312
|
contextWindow: number;
|
|
8228
8313
|
maxTokens: number;
|
|
8229
8314
|
};
|
|
8230
|
-
readonly "mistralai/mistral-small-3.1-24b-instruct:free": {
|
|
8231
|
-
id: string;
|
|
8232
|
-
name: string;
|
|
8233
|
-
api: "openai-completions";
|
|
8234
|
-
provider: string;
|
|
8235
|
-
baseUrl: string;
|
|
8236
|
-
reasoning: false;
|
|
8237
|
-
input: ("image" | "text")[];
|
|
8238
|
-
cost: {
|
|
8239
|
-
input: number;
|
|
8240
|
-
output: number;
|
|
8241
|
-
cacheRead: number;
|
|
8242
|
-
cacheWrite: number;
|
|
8243
|
-
};
|
|
8244
|
-
contextWindow: number;
|
|
8245
|
-
maxTokens: number;
|
|
8246
|
-
};
|
|
8247
8315
|
readonly "mistralai/mistral-small-3.2-24b-instruct": {
|
|
8248
8316
|
id: string;
|
|
8249
8317
|
name: string;
|
|
@@ -9842,23 +9910,6 @@ export declare const MODELS: {
|
|
|
9842
9910
|
contextWindow: number;
|
|
9843
9911
|
maxTokens: number;
|
|
9844
9912
|
};
|
|
9845
|
-
readonly "qwen/qwen3-4b:free": {
|
|
9846
|
-
id: string;
|
|
9847
|
-
name: string;
|
|
9848
|
-
api: "openai-completions";
|
|
9849
|
-
provider: string;
|
|
9850
|
-
baseUrl: string;
|
|
9851
|
-
reasoning: true;
|
|
9852
|
-
input: "text"[];
|
|
9853
|
-
cost: {
|
|
9854
|
-
input: number;
|
|
9855
|
-
output: number;
|
|
9856
|
-
cacheRead: number;
|
|
9857
|
-
cacheWrite: number;
|
|
9858
|
-
};
|
|
9859
|
-
contextWindow: number;
|
|
9860
|
-
maxTokens: number;
|
|
9861
|
-
};
|
|
9862
9913
|
readonly "qwen/qwen3-8b": {
|
|
9863
9914
|
id: string;
|
|
9864
9915
|
name: string;
|
|
@@ -11697,6 +11748,23 @@ export declare const MODELS: {
|
|
|
11697
11748
|
contextWindow: number;
|
|
11698
11749
|
maxTokens: number;
|
|
11699
11750
|
};
|
|
11751
|
+
readonly "kwaipilot/kat-coder-pro-v2": {
|
|
11752
|
+
id: string;
|
|
11753
|
+
name: string;
|
|
11754
|
+
api: "anthropic-messages";
|
|
11755
|
+
provider: string;
|
|
11756
|
+
baseUrl: string;
|
|
11757
|
+
reasoning: true;
|
|
11758
|
+
input: "text"[];
|
|
11759
|
+
cost: {
|
|
11760
|
+
input: number;
|
|
11761
|
+
output: number;
|
|
11762
|
+
cacheRead: number;
|
|
11763
|
+
cacheWrite: number;
|
|
11764
|
+
};
|
|
11765
|
+
contextWindow: number;
|
|
11766
|
+
maxTokens: number;
|
|
11767
|
+
};
|
|
11700
11768
|
readonly "meituan/longcat-flash-chat": {
|
|
11701
11769
|
id: string;
|
|
11702
11770
|
name: string;
|
|
@@ -13397,23 +13465,6 @@ export declare const MODELS: {
|
|
|
13397
13465
|
contextWindow: number;
|
|
13398
13466
|
maxTokens: number;
|
|
13399
13467
|
};
|
|
13400
|
-
readonly "zai/glm-5": {
|
|
13401
|
-
id: string;
|
|
13402
|
-
name: string;
|
|
13403
|
-
api: "anthropic-messages";
|
|
13404
|
-
provider: string;
|
|
13405
|
-
baseUrl: string;
|
|
13406
|
-
reasoning: true;
|
|
13407
|
-
input: "text"[];
|
|
13408
|
-
cost: {
|
|
13409
|
-
input: number;
|
|
13410
|
-
output: number;
|
|
13411
|
-
cacheRead: number;
|
|
13412
|
-
cacheWrite: number;
|
|
13413
|
-
};
|
|
13414
|
-
contextWindow: number;
|
|
13415
|
-
maxTokens: number;
|
|
13416
|
-
};
|
|
13417
13468
|
readonly "zai/glm-5-turbo": {
|
|
13418
13469
|
id: string;
|
|
13419
13470
|
name: string;
|