@mariozechner/pi-ai 0.52.11 → 0.53.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +1 -1
- package/dist/models.generated.d.ts +263 -16
- package/dist/models.generated.d.ts.map +1 -1
- package/dist/models.generated.js +289 -48
- package/dist/models.generated.js.map +1 -1
- package/dist/providers/openai-codex-responses.d.ts.map +1 -1
- package/dist/providers/openai-codex-responses.js +348 -0
- package/dist/providers/openai-codex-responses.js.map +1 -1
- package/dist/types.d.ts +6 -0
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +1 -1
package/README.md
CHANGED
|
@@ -1086,7 +1086,7 @@ const response = await complete(model, {
|
|
|
1086
1086
|
|
|
1087
1087
|
### Provider Notes
|
|
1088
1088
|
|
|
1089
|
-
**OpenAI Codex**: Requires a ChatGPT Plus or Pro subscription. Provides access to GPT-5.x Codex models with extended context windows and reasoning capabilities. The library automatically handles session-based prompt caching when `sessionId` is provided in stream options.
|
|
1089
|
+
**OpenAI Codex**: Requires a ChatGPT Plus or Pro subscription. Provides access to GPT-5.x Codex models with extended context windows and reasoning capabilities. The library automatically handles session-based prompt caching when `sessionId` is provided in stream options. You can set `transport` in stream options to `"sse"`, `"websocket"`, or `"auto"` for Codex Responses transport selection. When using WebSocket with a `sessionId`, connections are reused per session and expire after 5 minutes of inactivity.
|
|
1090
1090
|
|
|
1091
1091
|
**Azure OpenAI (Responses)**: Uses the Responses API only. Set `AZURE_OPENAI_API_KEY` and either `AZURE_OPENAI_BASE_URL` or `AZURE_OPENAI_RESOURCE_NAME`. Use `AZURE_OPENAI_API_VERSION` (defaults to `v1`) to override the API version if needed. Deployment names are treated as model IDs by default, override with `azureDeploymentName` or `AZURE_OPENAI_DEPLOYMENT_NAME_MAP` using comma-separated `model-id=deployment` pairs (for example `gpt-4o-mini=my-deployment,gpt-4o=prod`). Legacy deployment-based URLs are intentionally unsupported.
|
|
1092
1092
|
|
|
@@ -425,6 +425,23 @@ export declare const MODELS: {
|
|
|
425
425
|
contextWindow: number;
|
|
426
426
|
maxTokens: number;
|
|
427
427
|
};
|
|
428
|
+
readonly "deepseek.v3.2-v1:0": {
|
|
429
|
+
id: string;
|
|
430
|
+
name: string;
|
|
431
|
+
api: "bedrock-converse-stream";
|
|
432
|
+
provider: string;
|
|
433
|
+
baseUrl: string;
|
|
434
|
+
reasoning: true;
|
|
435
|
+
input: "text"[];
|
|
436
|
+
cost: {
|
|
437
|
+
input: number;
|
|
438
|
+
output: number;
|
|
439
|
+
cacheRead: number;
|
|
440
|
+
cacheWrite: number;
|
|
441
|
+
};
|
|
442
|
+
contextWindow: number;
|
|
443
|
+
maxTokens: number;
|
|
444
|
+
};
|
|
428
445
|
readonly "eu.anthropic.claude-haiku-4-5-20251001-v1:0": {
|
|
429
446
|
id: string;
|
|
430
447
|
name: string;
|
|
@@ -1702,6 +1719,23 @@ export declare const MODELS: {
|
|
|
1702
1719
|
contextWindow: number;
|
|
1703
1720
|
maxTokens: number;
|
|
1704
1721
|
};
|
|
1722
|
+
readonly "claude-sonnet-4-6": {
|
|
1723
|
+
id: string;
|
|
1724
|
+
name: string;
|
|
1725
|
+
api: "anthropic-messages";
|
|
1726
|
+
provider: string;
|
|
1727
|
+
baseUrl: string;
|
|
1728
|
+
reasoning: true;
|
|
1729
|
+
input: ("image" | "text")[];
|
|
1730
|
+
cost: {
|
|
1731
|
+
input: number;
|
|
1732
|
+
output: number;
|
|
1733
|
+
cacheRead: number;
|
|
1734
|
+
cacheWrite: number;
|
|
1735
|
+
};
|
|
1736
|
+
contextWindow: number;
|
|
1737
|
+
maxTokens: number;
|
|
1738
|
+
};
|
|
1705
1739
|
};
|
|
1706
1740
|
readonly "azure-openai-responses": {
|
|
1707
1741
|
readonly "codex-mini-latest": {
|
|
@@ -3890,6 +3924,26 @@ export declare const MODELS: {
|
|
|
3890
3924
|
contextWindow: number;
|
|
3891
3925
|
maxTokens: number;
|
|
3892
3926
|
};
|
|
3927
|
+
readonly "MiniMaxAI/MiniMax-M2.5": {
|
|
3928
|
+
id: string;
|
|
3929
|
+
name: string;
|
|
3930
|
+
api: "openai-completions";
|
|
3931
|
+
provider: string;
|
|
3932
|
+
baseUrl: string;
|
|
3933
|
+
compat: {
|
|
3934
|
+
supportsDeveloperRole: false;
|
|
3935
|
+
};
|
|
3936
|
+
reasoning: true;
|
|
3937
|
+
input: "text"[];
|
|
3938
|
+
cost: {
|
|
3939
|
+
input: number;
|
|
3940
|
+
output: number;
|
|
3941
|
+
cacheRead: number;
|
|
3942
|
+
cacheWrite: number;
|
|
3943
|
+
};
|
|
3944
|
+
contextWindow: number;
|
|
3945
|
+
maxTokens: number;
|
|
3946
|
+
};
|
|
3893
3947
|
readonly "Qwen/Qwen3-235B-A22B-Thinking-2507": {
|
|
3894
3948
|
id: string;
|
|
3895
3949
|
name: string;
|
|
@@ -3930,6 +3984,26 @@ export declare const MODELS: {
|
|
|
3930
3984
|
contextWindow: number;
|
|
3931
3985
|
maxTokens: number;
|
|
3932
3986
|
};
|
|
3987
|
+
readonly "Qwen/Qwen3-Coder-Next": {
|
|
3988
|
+
id: string;
|
|
3989
|
+
name: string;
|
|
3990
|
+
api: "openai-completions";
|
|
3991
|
+
provider: string;
|
|
3992
|
+
baseUrl: string;
|
|
3993
|
+
compat: {
|
|
3994
|
+
supportsDeveloperRole: false;
|
|
3995
|
+
};
|
|
3996
|
+
reasoning: false;
|
|
3997
|
+
input: "text"[];
|
|
3998
|
+
cost: {
|
|
3999
|
+
input: number;
|
|
4000
|
+
output: number;
|
|
4001
|
+
cacheRead: number;
|
|
4002
|
+
cacheWrite: number;
|
|
4003
|
+
};
|
|
4004
|
+
contextWindow: number;
|
|
4005
|
+
maxTokens: number;
|
|
4006
|
+
};
|
|
3933
4007
|
readonly "Qwen/Qwen3-Next-80B-A3B-Instruct": {
|
|
3934
4008
|
id: string;
|
|
3935
4009
|
name: string;
|
|
@@ -3970,6 +4044,26 @@ export declare const MODELS: {
|
|
|
3970
4044
|
contextWindow: number;
|
|
3971
4045
|
maxTokens: number;
|
|
3972
4046
|
};
|
|
4047
|
+
readonly "Qwen/Qwen3.5-397B-A17B": {
|
|
4048
|
+
id: string;
|
|
4049
|
+
name: string;
|
|
4050
|
+
api: "openai-completions";
|
|
4051
|
+
provider: string;
|
|
4052
|
+
baseUrl: string;
|
|
4053
|
+
compat: {
|
|
4054
|
+
supportsDeveloperRole: false;
|
|
4055
|
+
};
|
|
4056
|
+
reasoning: true;
|
|
4057
|
+
input: ("image" | "text")[];
|
|
4058
|
+
cost: {
|
|
4059
|
+
input: number;
|
|
4060
|
+
output: number;
|
|
4061
|
+
cacheRead: number;
|
|
4062
|
+
cacheWrite: number;
|
|
4063
|
+
};
|
|
4064
|
+
contextWindow: number;
|
|
4065
|
+
maxTokens: number;
|
|
4066
|
+
};
|
|
3973
4067
|
readonly "XiaomiMiMo/MiMo-V2-Flash": {
|
|
3974
4068
|
id: string;
|
|
3975
4069
|
name: string;
|
|
@@ -4259,6 +4353,23 @@ export declare const MODELS: {
|
|
|
4259
4353
|
contextWindow: number;
|
|
4260
4354
|
maxTokens: number;
|
|
4261
4355
|
};
|
|
4356
|
+
readonly "MiniMax-M2.5-highspeed": {
|
|
4357
|
+
id: string;
|
|
4358
|
+
name: string;
|
|
4359
|
+
api: "anthropic-messages";
|
|
4360
|
+
provider: string;
|
|
4361
|
+
baseUrl: string;
|
|
4362
|
+
reasoning: true;
|
|
4363
|
+
input: "text"[];
|
|
4364
|
+
cost: {
|
|
4365
|
+
input: number;
|
|
4366
|
+
output: number;
|
|
4367
|
+
cacheRead: number;
|
|
4368
|
+
cacheWrite: number;
|
|
4369
|
+
};
|
|
4370
|
+
contextWindow: number;
|
|
4371
|
+
maxTokens: number;
|
|
4372
|
+
};
|
|
4262
4373
|
};
|
|
4263
4374
|
readonly "minimax-cn": {
|
|
4264
4375
|
readonly "MiniMax-M2": {
|
|
@@ -4312,6 +4423,23 @@ export declare const MODELS: {
|
|
|
4312
4423
|
contextWindow: number;
|
|
4313
4424
|
maxTokens: number;
|
|
4314
4425
|
};
|
|
4426
|
+
readonly "MiniMax-M2.5-highspeed": {
|
|
4427
|
+
id: string;
|
|
4428
|
+
name: string;
|
|
4429
|
+
api: "anthropic-messages";
|
|
4430
|
+
provider: string;
|
|
4431
|
+
baseUrl: string;
|
|
4432
|
+
reasoning: true;
|
|
4433
|
+
input: "text"[];
|
|
4434
|
+
cost: {
|
|
4435
|
+
input: number;
|
|
4436
|
+
output: number;
|
|
4437
|
+
cacheRead: number;
|
|
4438
|
+
cacheWrite: number;
|
|
4439
|
+
};
|
|
4440
|
+
contextWindow: number;
|
|
4441
|
+
maxTokens: number;
|
|
4442
|
+
};
|
|
4315
4443
|
};
|
|
4316
4444
|
readonly mistral: {
|
|
4317
4445
|
readonly "codestral-latest": {
|
|
@@ -5680,6 +5808,40 @@ export declare const MODELS: {
|
|
|
5680
5808
|
contextWindow: number;
|
|
5681
5809
|
maxTokens: number;
|
|
5682
5810
|
};
|
|
5811
|
+
readonly "glm-5": {
|
|
5812
|
+
id: string;
|
|
5813
|
+
name: string;
|
|
5814
|
+
api: "openai-completions";
|
|
5815
|
+
provider: string;
|
|
5816
|
+
baseUrl: string;
|
|
5817
|
+
reasoning: true;
|
|
5818
|
+
input: "text"[];
|
|
5819
|
+
cost: {
|
|
5820
|
+
input: number;
|
|
5821
|
+
output: number;
|
|
5822
|
+
cacheRead: number;
|
|
5823
|
+
cacheWrite: number;
|
|
5824
|
+
};
|
|
5825
|
+
contextWindow: number;
|
|
5826
|
+
maxTokens: number;
|
|
5827
|
+
};
|
|
5828
|
+
readonly "glm-5-free": {
|
|
5829
|
+
id: string;
|
|
5830
|
+
name: string;
|
|
5831
|
+
api: "openai-completions";
|
|
5832
|
+
provider: string;
|
|
5833
|
+
baseUrl: string;
|
|
5834
|
+
reasoning: true;
|
|
5835
|
+
input: "text"[];
|
|
5836
|
+
cost: {
|
|
5837
|
+
input: number;
|
|
5838
|
+
output: number;
|
|
5839
|
+
cacheRead: number;
|
|
5840
|
+
cacheWrite: number;
|
|
5841
|
+
};
|
|
5842
|
+
contextWindow: number;
|
|
5843
|
+
maxTokens: number;
|
|
5844
|
+
};
|
|
5683
5845
|
readonly "gpt-5": {
|
|
5684
5846
|
id: string;
|
|
5685
5847
|
name: string;
|
|
@@ -5918,6 +6080,23 @@ export declare const MODELS: {
|
|
|
5918
6080
|
contextWindow: number;
|
|
5919
6081
|
maxTokens: number;
|
|
5920
6082
|
};
|
|
6083
|
+
readonly "minimax-m2.5": {
|
|
6084
|
+
id: string;
|
|
6085
|
+
name: string;
|
|
6086
|
+
api: "openai-completions";
|
|
6087
|
+
provider: string;
|
|
6088
|
+
baseUrl: string;
|
|
6089
|
+
reasoning: true;
|
|
6090
|
+
input: "text"[];
|
|
6091
|
+
cost: {
|
|
6092
|
+
input: number;
|
|
6093
|
+
output: number;
|
|
6094
|
+
cacheRead: number;
|
|
6095
|
+
cacheWrite: number;
|
|
6096
|
+
};
|
|
6097
|
+
contextWindow: number;
|
|
6098
|
+
maxTokens: number;
|
|
6099
|
+
};
|
|
5921
6100
|
readonly "minimax-m2.5-free": {
|
|
5922
6101
|
id: string;
|
|
5923
6102
|
name: string;
|
|
@@ -6277,6 +6456,23 @@ export declare const MODELS: {
|
|
|
6277
6456
|
contextWindow: number;
|
|
6278
6457
|
maxTokens: number;
|
|
6279
6458
|
};
|
|
6459
|
+
readonly "anthropic/claude-sonnet-4.6": {
|
|
6460
|
+
id: string;
|
|
6461
|
+
name: string;
|
|
6462
|
+
api: "openai-completions";
|
|
6463
|
+
provider: string;
|
|
6464
|
+
baseUrl: string;
|
|
6465
|
+
reasoning: true;
|
|
6466
|
+
input: ("image" | "text")[];
|
|
6467
|
+
cost: {
|
|
6468
|
+
input: number;
|
|
6469
|
+
output: number;
|
|
6470
|
+
cacheRead: number;
|
|
6471
|
+
cacheWrite: number;
|
|
6472
|
+
};
|
|
6473
|
+
contextWindow: number;
|
|
6474
|
+
maxTokens: number;
|
|
6475
|
+
};
|
|
6280
6476
|
readonly "arcee-ai/trinity-large-preview:free": {
|
|
6281
6477
|
id: string;
|
|
6282
6478
|
name: string;
|
|
@@ -9320,14 +9516,14 @@ export declare const MODELS: {
|
|
|
9320
9516
|
contextWindow: number;
|
|
9321
9517
|
maxTokens: number;
|
|
9322
9518
|
};
|
|
9323
|
-
readonly "qwen/
|
|
9519
|
+
readonly "qwen/qwen3.5-397b-a17b": {
|
|
9324
9520
|
id: string;
|
|
9325
9521
|
name: string;
|
|
9326
9522
|
api: "openai-completions";
|
|
9327
9523
|
provider: string;
|
|
9328
9524
|
baseUrl: string;
|
|
9329
9525
|
reasoning: true;
|
|
9330
|
-
input: "text"[];
|
|
9526
|
+
input: ("image" | "text")[];
|
|
9331
9527
|
cost: {
|
|
9332
9528
|
input: number;
|
|
9333
9529
|
output: number;
|
|
@@ -9337,13 +9533,30 @@ export declare const MODELS: {
|
|
|
9337
9533
|
contextWindow: number;
|
|
9338
9534
|
maxTokens: number;
|
|
9339
9535
|
};
|
|
9340
|
-
readonly "
|
|
9536
|
+
readonly "qwen/qwen3.5-plus-02-15": {
|
|
9341
9537
|
id: string;
|
|
9342
9538
|
name: string;
|
|
9343
9539
|
api: "openai-completions";
|
|
9344
9540
|
provider: string;
|
|
9345
9541
|
baseUrl: string;
|
|
9346
|
-
reasoning:
|
|
9542
|
+
reasoning: true;
|
|
9543
|
+
input: ("image" | "text")[];
|
|
9544
|
+
cost: {
|
|
9545
|
+
input: number;
|
|
9546
|
+
output: number;
|
|
9547
|
+
cacheRead: number;
|
|
9548
|
+
cacheWrite: number;
|
|
9549
|
+
};
|
|
9550
|
+
contextWindow: number;
|
|
9551
|
+
maxTokens: number;
|
|
9552
|
+
};
|
|
9553
|
+
readonly "qwen/qwq-32b": {
|
|
9554
|
+
id: string;
|
|
9555
|
+
name: string;
|
|
9556
|
+
api: "openai-completions";
|
|
9557
|
+
provider: string;
|
|
9558
|
+
baseUrl: string;
|
|
9559
|
+
reasoning: true;
|
|
9347
9560
|
input: "text"[];
|
|
9348
9561
|
cost: {
|
|
9349
9562
|
input: number;
|
|
@@ -9354,7 +9567,7 @@ export declare const MODELS: {
|
|
|
9354
9567
|
contextWindow: number;
|
|
9355
9568
|
maxTokens: number;
|
|
9356
9569
|
};
|
|
9357
|
-
readonly "
|
|
9570
|
+
readonly "relace/relace-search": {
|
|
9358
9571
|
id: string;
|
|
9359
9572
|
name: string;
|
|
9360
9573
|
api: "openai-completions";
|
|
@@ -9371,7 +9584,7 @@ export declare const MODELS: {
|
|
|
9371
9584
|
contextWindow: number;
|
|
9372
9585
|
maxTokens: number;
|
|
9373
9586
|
};
|
|
9374
|
-
readonly "sao10k/l3
|
|
9587
|
+
readonly "sao10k/l3-euryale-70b": {
|
|
9375
9588
|
id: string;
|
|
9376
9589
|
name: string;
|
|
9377
9590
|
api: "openai-completions";
|
|
@@ -9388,13 +9601,13 @@ export declare const MODELS: {
|
|
|
9388
9601
|
contextWindow: number;
|
|
9389
9602
|
maxTokens: number;
|
|
9390
9603
|
};
|
|
9391
|
-
readonly "
|
|
9604
|
+
readonly "sao10k/l3.1-euryale-70b": {
|
|
9392
9605
|
id: string;
|
|
9393
9606
|
name: string;
|
|
9394
9607
|
api: "openai-completions";
|
|
9395
9608
|
provider: string;
|
|
9396
9609
|
baseUrl: string;
|
|
9397
|
-
reasoning:
|
|
9610
|
+
reasoning: false;
|
|
9398
9611
|
input: "text"[];
|
|
9399
9612
|
cost: {
|
|
9400
9613
|
input: number;
|
|
@@ -9405,7 +9618,7 @@ export declare const MODELS: {
|
|
|
9405
9618
|
contextWindow: number;
|
|
9406
9619
|
maxTokens: number;
|
|
9407
9620
|
};
|
|
9408
|
-
readonly "stepfun/step-3.5-flash
|
|
9621
|
+
readonly "stepfun/step-3.5-flash": {
|
|
9409
9622
|
id: string;
|
|
9410
9623
|
name: string;
|
|
9411
9624
|
api: "openai-completions";
|
|
@@ -9422,13 +9635,13 @@ export declare const MODELS: {
|
|
|
9422
9635
|
contextWindow: number;
|
|
9423
9636
|
maxTokens: number;
|
|
9424
9637
|
};
|
|
9425
|
-
readonly "
|
|
9638
|
+
readonly "stepfun/step-3.5-flash:free": {
|
|
9426
9639
|
id: string;
|
|
9427
9640
|
name: string;
|
|
9428
9641
|
api: "openai-completions";
|
|
9429
9642
|
provider: string;
|
|
9430
9643
|
baseUrl: string;
|
|
9431
|
-
reasoning:
|
|
9644
|
+
reasoning: true;
|
|
9432
9645
|
input: "text"[];
|
|
9433
9646
|
cost: {
|
|
9434
9647
|
input: number;
|
|
@@ -9439,7 +9652,7 @@ export declare const MODELS: {
|
|
|
9439
9652
|
contextWindow: number;
|
|
9440
9653
|
maxTokens: number;
|
|
9441
9654
|
};
|
|
9442
|
-
readonly "thedrummer/
|
|
9655
|
+
readonly "thedrummer/rocinante-12b": {
|
|
9443
9656
|
id: string;
|
|
9444
9657
|
name: string;
|
|
9445
9658
|
api: "openai-completions";
|
|
@@ -9456,13 +9669,13 @@ export declare const MODELS: {
|
|
|
9456
9669
|
contextWindow: number;
|
|
9457
9670
|
maxTokens: number;
|
|
9458
9671
|
};
|
|
9459
|
-
readonly "
|
|
9672
|
+
readonly "thedrummer/unslopnemo-12b": {
|
|
9460
9673
|
id: string;
|
|
9461
9674
|
name: string;
|
|
9462
9675
|
api: "openai-completions";
|
|
9463
9676
|
provider: string;
|
|
9464
9677
|
baseUrl: string;
|
|
9465
|
-
reasoning:
|
|
9678
|
+
reasoning: false;
|
|
9466
9679
|
input: "text"[];
|
|
9467
9680
|
cost: {
|
|
9468
9681
|
input: number;
|
|
@@ -9473,7 +9686,7 @@ export declare const MODELS: {
|
|
|
9473
9686
|
contextWindow: number;
|
|
9474
9687
|
maxTokens: number;
|
|
9475
9688
|
};
|
|
9476
|
-
readonly "tngtech/
|
|
9689
|
+
readonly "tngtech/deepseek-r1t2-chimera": {
|
|
9477
9690
|
id: string;
|
|
9478
9691
|
name: string;
|
|
9479
9692
|
api: "openai-completions";
|
|
@@ -9490,7 +9703,7 @@ export declare const MODELS: {
|
|
|
9490
9703
|
contextWindow: number;
|
|
9491
9704
|
maxTokens: number;
|
|
9492
9705
|
};
|
|
9493
|
-
readonly "tngtech/tng-r1t-chimera
|
|
9706
|
+
readonly "tngtech/tng-r1t-chimera": {
|
|
9494
9707
|
id: string;
|
|
9495
9708
|
name: string;
|
|
9496
9709
|
api: "openai-completions";
|
|
@@ -10070,6 +10283,23 @@ export declare const MODELS: {
|
|
|
10070
10283
|
contextWindow: number;
|
|
10071
10284
|
maxTokens: number;
|
|
10072
10285
|
};
|
|
10286
|
+
readonly "alibaba/qwen3.5-plus": {
|
|
10287
|
+
id: string;
|
|
10288
|
+
name: string;
|
|
10289
|
+
api: "anthropic-messages";
|
|
10290
|
+
provider: string;
|
|
10291
|
+
baseUrl: string;
|
|
10292
|
+
reasoning: true;
|
|
10293
|
+
input: ("image" | "text")[];
|
|
10294
|
+
cost: {
|
|
10295
|
+
input: number;
|
|
10296
|
+
output: number;
|
|
10297
|
+
cacheRead: number;
|
|
10298
|
+
cacheWrite: number;
|
|
10299
|
+
};
|
|
10300
|
+
contextWindow: number;
|
|
10301
|
+
maxTokens: number;
|
|
10302
|
+
};
|
|
10073
10303
|
readonly "anthropic/claude-3-haiku": {
|
|
10074
10304
|
id: string;
|
|
10075
10305
|
name: string;
|
|
@@ -10274,6 +10504,23 @@ export declare const MODELS: {
|
|
|
10274
10504
|
contextWindow: number;
|
|
10275
10505
|
maxTokens: number;
|
|
10276
10506
|
};
|
|
10507
|
+
readonly "anthropic/claude-sonnet-4.6": {
|
|
10508
|
+
id: string;
|
|
10509
|
+
name: string;
|
|
10510
|
+
api: "anthropic-messages";
|
|
10511
|
+
provider: string;
|
|
10512
|
+
baseUrl: string;
|
|
10513
|
+
reasoning: true;
|
|
10514
|
+
input: ("image" | "text")[];
|
|
10515
|
+
cost: {
|
|
10516
|
+
input: number;
|
|
10517
|
+
output: number;
|
|
10518
|
+
cacheRead: number;
|
|
10519
|
+
cacheWrite: number;
|
|
10520
|
+
};
|
|
10521
|
+
contextWindow: number;
|
|
10522
|
+
maxTokens: number;
|
|
10523
|
+
};
|
|
10277
10524
|
readonly "arcee-ai/trinity-large-preview": {
|
|
10278
10525
|
id: string;
|
|
10279
10526
|
name: string;
|