@mariozechner/pi-ai 0.35.0 → 0.36.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (50) hide show
  1. package/dist/cli.d.ts.map +1 -1
  2. package/dist/cli.js +17 -1
  3. package/dist/cli.js.map +1 -1
  4. package/dist/models.generated.d.ts +640 -43
  5. package/dist/models.generated.d.ts.map +1 -1
  6. package/dist/models.generated.js +673 -76
  7. package/dist/models.generated.js.map +1 -1
  8. package/dist/providers/openai-codex/constants.d.ts +21 -0
  9. package/dist/providers/openai-codex/constants.d.ts.map +1 -0
  10. package/dist/providers/openai-codex/constants.js +21 -0
  11. package/dist/providers/openai-codex/constants.js.map +1 -0
  12. package/dist/providers/openai-codex/prompts/codex-instructions.md +105 -0
  13. package/dist/providers/openai-codex/prompts/codex.d.ts +11 -0
  14. package/dist/providers/openai-codex/prompts/codex.d.ts.map +1 -0
  15. package/dist/providers/openai-codex/prompts/codex.js +184 -0
  16. package/dist/providers/openai-codex/prompts/codex.js.map +1 -0
  17. package/dist/providers/openai-codex/prompts/pi-codex-bridge.d.ts +6 -0
  18. package/dist/providers/openai-codex/prompts/pi-codex-bridge.d.ts.map +1 -0
  19. package/dist/providers/openai-codex/prompts/pi-codex-bridge.js +48 -0
  20. package/dist/providers/openai-codex/prompts/pi-codex-bridge.js.map +1 -0
  21. package/dist/providers/openai-codex/request-transformer.d.ts +41 -0
  22. package/dist/providers/openai-codex/request-transformer.d.ts.map +1 -0
  23. package/dist/providers/openai-codex/request-transformer.js +242 -0
  24. package/dist/providers/openai-codex/request-transformer.js.map +1 -0
  25. package/dist/providers/openai-codex/response-handler.d.ts +19 -0
  26. package/dist/providers/openai-codex/response-handler.d.ts.map +1 -0
  27. package/dist/providers/openai-codex/response-handler.js +107 -0
  28. package/dist/providers/openai-codex/response-handler.js.map +1 -0
  29. package/dist/providers/openai-codex-responses.d.ts +10 -0
  30. package/dist/providers/openai-codex-responses.d.ts.map +1 -0
  31. package/dist/providers/openai-codex-responses.js +528 -0
  32. package/dist/providers/openai-codex-responses.js.map +1 -0
  33. package/dist/stream.d.ts.map +1 -1
  34. package/dist/stream.js +27 -1
  35. package/dist/stream.js.map +1 -1
  36. package/dist/types.d.ts +4 -2
  37. package/dist/types.d.ts.map +1 -1
  38. package/dist/types.js.map +1 -1
  39. package/dist/utils/oauth/index.d.ts +1 -0
  40. package/dist/utils/oauth/index.d.ts.map +1 -1
  41. package/dist/utils/oauth/index.js +11 -0
  42. package/dist/utils/oauth/index.js.map +1 -1
  43. package/dist/utils/oauth/openai-codex.d.ts +20 -0
  44. package/dist/utils/oauth/openai-codex.d.ts.map +1 -0
  45. package/dist/utils/oauth/openai-codex.js +278 -0
  46. package/dist/utils/oauth/openai-codex.js.map +1 -0
  47. package/dist/utils/oauth/types.d.ts +2 -1
  48. package/dist/utils/oauth/types.d.ts.map +1 -1
  49. package/dist/utils/oauth/types.js.map +1 -1
  50. package/package.json +2 -2
@@ -2903,15 +2903,423 @@ export declare const MODELS: {
2903
2903
  maxTokens: number;
2904
2904
  };
2905
2905
  };
2906
- readonly openrouter: {
2907
- readonly "ai21/jamba-large-1.7": {
2906
+ readonly "openai-codex": {
2907
+ readonly "codex-mini-latest": {
2908
+ id: string;
2909
+ name: string;
2910
+ api: "openai-codex-responses";
2911
+ provider: string;
2912
+ baseUrl: string;
2913
+ reasoning: true;
2914
+ input: ("image" | "text")[];
2915
+ cost: {
2916
+ input: number;
2917
+ output: number;
2918
+ cacheRead: number;
2919
+ cacheWrite: number;
2920
+ };
2921
+ contextWindow: number;
2922
+ maxTokens: number;
2923
+ };
2924
+ readonly "gpt-5": {
2925
+ id: string;
2926
+ name: string;
2927
+ api: "openai-codex-responses";
2928
+ provider: string;
2929
+ baseUrl: string;
2930
+ reasoning: true;
2931
+ input: ("image" | "text")[];
2932
+ cost: {
2933
+ input: number;
2934
+ output: number;
2935
+ cacheRead: number;
2936
+ cacheWrite: number;
2937
+ };
2938
+ contextWindow: number;
2939
+ maxTokens: number;
2940
+ };
2941
+ readonly "gpt-5-codex": {
2942
+ id: string;
2943
+ name: string;
2944
+ api: "openai-codex-responses";
2945
+ provider: string;
2946
+ baseUrl: string;
2947
+ reasoning: true;
2948
+ input: ("image" | "text")[];
2949
+ cost: {
2950
+ input: number;
2951
+ output: number;
2952
+ cacheRead: number;
2953
+ cacheWrite: number;
2954
+ };
2955
+ contextWindow: number;
2956
+ maxTokens: number;
2957
+ };
2958
+ readonly "gpt-5-codex-mini": {
2959
+ id: string;
2960
+ name: string;
2961
+ api: "openai-codex-responses";
2962
+ provider: string;
2963
+ baseUrl: string;
2964
+ reasoning: true;
2965
+ input: ("image" | "text")[];
2966
+ cost: {
2967
+ input: number;
2968
+ output: number;
2969
+ cacheRead: number;
2970
+ cacheWrite: number;
2971
+ };
2972
+ contextWindow: number;
2973
+ maxTokens: number;
2974
+ };
2975
+ readonly "gpt-5-codex-mini-high": {
2976
+ id: string;
2977
+ name: string;
2978
+ api: "openai-codex-responses";
2979
+ provider: string;
2980
+ baseUrl: string;
2981
+ reasoning: true;
2982
+ input: ("image" | "text")[];
2983
+ cost: {
2984
+ input: number;
2985
+ output: number;
2986
+ cacheRead: number;
2987
+ cacheWrite: number;
2988
+ };
2989
+ contextWindow: number;
2990
+ maxTokens: number;
2991
+ };
2992
+ readonly "gpt-5-codex-mini-medium": {
2993
+ id: string;
2994
+ name: string;
2995
+ api: "openai-codex-responses";
2996
+ provider: string;
2997
+ baseUrl: string;
2998
+ reasoning: true;
2999
+ input: ("image" | "text")[];
3000
+ cost: {
3001
+ input: number;
3002
+ output: number;
3003
+ cacheRead: number;
3004
+ cacheWrite: number;
3005
+ };
3006
+ contextWindow: number;
3007
+ maxTokens: number;
3008
+ };
3009
+ readonly "gpt-5-mini": {
3010
+ id: string;
3011
+ name: string;
3012
+ api: "openai-codex-responses";
3013
+ provider: string;
3014
+ baseUrl: string;
3015
+ reasoning: true;
3016
+ input: ("image" | "text")[];
3017
+ cost: {
3018
+ input: number;
3019
+ output: number;
3020
+ cacheRead: number;
3021
+ cacheWrite: number;
3022
+ };
3023
+ contextWindow: number;
3024
+ maxTokens: number;
3025
+ };
3026
+ readonly "gpt-5-nano": {
3027
+ id: string;
3028
+ name: string;
3029
+ api: "openai-codex-responses";
3030
+ provider: string;
3031
+ baseUrl: string;
3032
+ reasoning: true;
3033
+ input: ("image" | "text")[];
3034
+ cost: {
3035
+ input: number;
3036
+ output: number;
3037
+ cacheRead: number;
3038
+ cacheWrite: number;
3039
+ };
3040
+ contextWindow: number;
3041
+ maxTokens: number;
3042
+ };
3043
+ readonly "gpt-5.1": {
3044
+ id: string;
3045
+ name: string;
3046
+ api: "openai-codex-responses";
3047
+ provider: string;
3048
+ baseUrl: string;
3049
+ reasoning: true;
3050
+ input: ("image" | "text")[];
3051
+ cost: {
3052
+ input: number;
3053
+ output: number;
3054
+ cacheRead: number;
3055
+ cacheWrite: number;
3056
+ };
3057
+ contextWindow: number;
3058
+ maxTokens: number;
3059
+ };
3060
+ readonly "gpt-5.1-chat-latest": {
3061
+ id: string;
3062
+ name: string;
3063
+ api: "openai-codex-responses";
3064
+ provider: string;
3065
+ baseUrl: string;
3066
+ reasoning: true;
3067
+ input: ("image" | "text")[];
3068
+ cost: {
3069
+ input: number;
3070
+ output: number;
3071
+ cacheRead: number;
3072
+ cacheWrite: number;
3073
+ };
3074
+ contextWindow: number;
3075
+ maxTokens: number;
3076
+ };
3077
+ readonly "gpt-5.1-codex": {
3078
+ id: string;
3079
+ name: string;
3080
+ api: "openai-codex-responses";
3081
+ provider: string;
3082
+ baseUrl: string;
3083
+ reasoning: true;
3084
+ input: ("image" | "text")[];
3085
+ cost: {
3086
+ input: number;
3087
+ output: number;
3088
+ cacheRead: number;
3089
+ cacheWrite: number;
3090
+ };
3091
+ contextWindow: number;
3092
+ maxTokens: number;
3093
+ };
3094
+ readonly "gpt-5.1-codex-high": {
3095
+ id: string;
3096
+ name: string;
3097
+ api: "openai-codex-responses";
3098
+ provider: string;
3099
+ baseUrl: string;
3100
+ reasoning: true;
3101
+ input: ("image" | "text")[];
3102
+ cost: {
3103
+ input: number;
3104
+ output: number;
3105
+ cacheRead: number;
3106
+ cacheWrite: number;
3107
+ };
3108
+ contextWindow: number;
3109
+ maxTokens: number;
3110
+ };
3111
+ readonly "gpt-5.1-codex-low": {
3112
+ id: string;
3113
+ name: string;
3114
+ api: "openai-codex-responses";
3115
+ provider: string;
3116
+ baseUrl: string;
3117
+ reasoning: true;
3118
+ input: ("image" | "text")[];
3119
+ cost: {
3120
+ input: number;
3121
+ output: number;
3122
+ cacheRead: number;
3123
+ cacheWrite: number;
3124
+ };
3125
+ contextWindow: number;
3126
+ maxTokens: number;
3127
+ };
3128
+ readonly "gpt-5.1-codex-max": {
3129
+ id: string;
3130
+ name: string;
3131
+ api: "openai-codex-responses";
3132
+ provider: string;
3133
+ baseUrl: string;
3134
+ reasoning: true;
3135
+ input: ("image" | "text")[];
3136
+ cost: {
3137
+ input: number;
3138
+ output: number;
3139
+ cacheRead: number;
3140
+ cacheWrite: number;
3141
+ };
3142
+ contextWindow: number;
3143
+ maxTokens: number;
3144
+ };
3145
+ readonly "gpt-5.1-codex-max-high": {
3146
+ id: string;
3147
+ name: string;
3148
+ api: "openai-codex-responses";
3149
+ provider: string;
3150
+ baseUrl: string;
3151
+ reasoning: true;
3152
+ input: ("image" | "text")[];
3153
+ cost: {
3154
+ input: number;
3155
+ output: number;
3156
+ cacheRead: number;
3157
+ cacheWrite: number;
3158
+ };
3159
+ contextWindow: number;
3160
+ maxTokens: number;
3161
+ };
3162
+ readonly "gpt-5.1-codex-max-low": {
3163
+ id: string;
3164
+ name: string;
3165
+ api: "openai-codex-responses";
3166
+ provider: string;
3167
+ baseUrl: string;
3168
+ reasoning: true;
3169
+ input: ("image" | "text")[];
3170
+ cost: {
3171
+ input: number;
3172
+ output: number;
3173
+ cacheRead: number;
3174
+ cacheWrite: number;
3175
+ };
3176
+ contextWindow: number;
3177
+ maxTokens: number;
3178
+ };
3179
+ readonly "gpt-5.1-codex-max-medium": {
3180
+ id: string;
3181
+ name: string;
3182
+ api: "openai-codex-responses";
3183
+ provider: string;
3184
+ baseUrl: string;
3185
+ reasoning: true;
3186
+ input: ("image" | "text")[];
3187
+ cost: {
3188
+ input: number;
3189
+ output: number;
3190
+ cacheRead: number;
3191
+ cacheWrite: number;
3192
+ };
3193
+ contextWindow: number;
3194
+ maxTokens: number;
3195
+ };
3196
+ readonly "gpt-5.1-codex-max-xhigh": {
3197
+ id: string;
3198
+ name: string;
3199
+ api: "openai-codex-responses";
3200
+ provider: string;
3201
+ baseUrl: string;
3202
+ reasoning: true;
3203
+ input: ("image" | "text")[];
3204
+ cost: {
3205
+ input: number;
3206
+ output: number;
3207
+ cacheRead: number;
3208
+ cacheWrite: number;
3209
+ };
3210
+ contextWindow: number;
3211
+ maxTokens: number;
3212
+ };
3213
+ readonly "gpt-5.1-codex-medium": {
3214
+ id: string;
3215
+ name: string;
3216
+ api: "openai-codex-responses";
3217
+ provider: string;
3218
+ baseUrl: string;
3219
+ reasoning: true;
3220
+ input: ("image" | "text")[];
3221
+ cost: {
3222
+ input: number;
3223
+ output: number;
3224
+ cacheRead: number;
3225
+ cacheWrite: number;
3226
+ };
3227
+ contextWindow: number;
3228
+ maxTokens: number;
3229
+ };
3230
+ readonly "gpt-5.1-codex-mini": {
3231
+ id: string;
3232
+ name: string;
3233
+ api: "openai-codex-responses";
3234
+ provider: string;
3235
+ baseUrl: string;
3236
+ reasoning: true;
3237
+ input: ("image" | "text")[];
3238
+ cost: {
3239
+ input: number;
3240
+ output: number;
3241
+ cacheRead: number;
3242
+ cacheWrite: number;
3243
+ };
3244
+ contextWindow: number;
3245
+ maxTokens: number;
3246
+ };
3247
+ readonly "gpt-5.1-codex-mini-high": {
3248
+ id: string;
3249
+ name: string;
3250
+ api: "openai-codex-responses";
3251
+ provider: string;
3252
+ baseUrl: string;
3253
+ reasoning: true;
3254
+ input: ("image" | "text")[];
3255
+ cost: {
3256
+ input: number;
3257
+ output: number;
3258
+ cacheRead: number;
3259
+ cacheWrite: number;
3260
+ };
3261
+ contextWindow: number;
3262
+ maxTokens: number;
3263
+ };
3264
+ readonly "gpt-5.1-codex-mini-medium": {
3265
+ id: string;
3266
+ name: string;
3267
+ api: "openai-codex-responses";
3268
+ provider: string;
3269
+ baseUrl: string;
3270
+ reasoning: true;
3271
+ input: ("image" | "text")[];
3272
+ cost: {
3273
+ input: number;
3274
+ output: number;
3275
+ cacheRead: number;
3276
+ cacheWrite: number;
3277
+ };
3278
+ contextWindow: number;
3279
+ maxTokens: number;
3280
+ };
3281
+ readonly "gpt-5.1-high": {
3282
+ id: string;
3283
+ name: string;
3284
+ api: "openai-codex-responses";
3285
+ provider: string;
3286
+ baseUrl: string;
3287
+ reasoning: true;
3288
+ input: ("image" | "text")[];
3289
+ cost: {
3290
+ input: number;
3291
+ output: number;
3292
+ cacheRead: number;
3293
+ cacheWrite: number;
3294
+ };
3295
+ contextWindow: number;
3296
+ maxTokens: number;
3297
+ };
3298
+ readonly "gpt-5.1-low": {
3299
+ id: string;
3300
+ name: string;
3301
+ api: "openai-codex-responses";
3302
+ provider: string;
3303
+ baseUrl: string;
3304
+ reasoning: true;
3305
+ input: ("image" | "text")[];
3306
+ cost: {
3307
+ input: number;
3308
+ output: number;
3309
+ cacheRead: number;
3310
+ cacheWrite: number;
3311
+ };
3312
+ contextWindow: number;
3313
+ maxTokens: number;
3314
+ };
3315
+ readonly "gpt-5.1-medium": {
2908
3316
  id: string;
2909
3317
  name: string;
2910
- api: "openai-completions";
3318
+ api: "openai-codex-responses";
2911
3319
  provider: string;
2912
3320
  baseUrl: string;
2913
- reasoning: false;
2914
- input: "text"[];
3321
+ reasoning: true;
3322
+ input: ("image" | "text")[];
2915
3323
  cost: {
2916
3324
  input: number;
2917
3325
  output: number;
@@ -2921,7 +3329,213 @@ export declare const MODELS: {
2921
3329
  contextWindow: number;
2922
3330
  maxTokens: number;
2923
3331
  };
2924
- readonly "ai21/jamba-mini-1.7": {
3332
+ readonly "gpt-5.1-none": {
3333
+ id: string;
3334
+ name: string;
3335
+ api: "openai-codex-responses";
3336
+ provider: string;
3337
+ baseUrl: string;
3338
+ reasoning: true;
3339
+ input: ("image" | "text")[];
3340
+ cost: {
3341
+ input: number;
3342
+ output: number;
3343
+ cacheRead: number;
3344
+ cacheWrite: number;
3345
+ };
3346
+ contextWindow: number;
3347
+ maxTokens: number;
3348
+ };
3349
+ readonly "gpt-5.2": {
3350
+ id: string;
3351
+ name: string;
3352
+ api: "openai-codex-responses";
3353
+ provider: string;
3354
+ baseUrl: string;
3355
+ reasoning: true;
3356
+ input: ("image" | "text")[];
3357
+ cost: {
3358
+ input: number;
3359
+ output: number;
3360
+ cacheRead: number;
3361
+ cacheWrite: number;
3362
+ };
3363
+ contextWindow: number;
3364
+ maxTokens: number;
3365
+ };
3366
+ readonly "gpt-5.2-codex": {
3367
+ id: string;
3368
+ name: string;
3369
+ api: "openai-codex-responses";
3370
+ provider: string;
3371
+ baseUrl: string;
3372
+ reasoning: true;
3373
+ input: ("image" | "text")[];
3374
+ cost: {
3375
+ input: number;
3376
+ output: number;
3377
+ cacheRead: number;
3378
+ cacheWrite: number;
3379
+ };
3380
+ contextWindow: number;
3381
+ maxTokens: number;
3382
+ };
3383
+ readonly "gpt-5.2-codex-high": {
3384
+ id: string;
3385
+ name: string;
3386
+ api: "openai-codex-responses";
3387
+ provider: string;
3388
+ baseUrl: string;
3389
+ reasoning: true;
3390
+ input: ("image" | "text")[];
3391
+ cost: {
3392
+ input: number;
3393
+ output: number;
3394
+ cacheRead: number;
3395
+ cacheWrite: number;
3396
+ };
3397
+ contextWindow: number;
3398
+ maxTokens: number;
3399
+ };
3400
+ readonly "gpt-5.2-codex-low": {
3401
+ id: string;
3402
+ name: string;
3403
+ api: "openai-codex-responses";
3404
+ provider: string;
3405
+ baseUrl: string;
3406
+ reasoning: true;
3407
+ input: ("image" | "text")[];
3408
+ cost: {
3409
+ input: number;
3410
+ output: number;
3411
+ cacheRead: number;
3412
+ cacheWrite: number;
3413
+ };
3414
+ contextWindow: number;
3415
+ maxTokens: number;
3416
+ };
3417
+ readonly "gpt-5.2-codex-medium": {
3418
+ id: string;
3419
+ name: string;
3420
+ api: "openai-codex-responses";
3421
+ provider: string;
3422
+ baseUrl: string;
3423
+ reasoning: true;
3424
+ input: ("image" | "text")[];
3425
+ cost: {
3426
+ input: number;
3427
+ output: number;
3428
+ cacheRead: number;
3429
+ cacheWrite: number;
3430
+ };
3431
+ contextWindow: number;
3432
+ maxTokens: number;
3433
+ };
3434
+ readonly "gpt-5.2-codex-xhigh": {
3435
+ id: string;
3436
+ name: string;
3437
+ api: "openai-codex-responses";
3438
+ provider: string;
3439
+ baseUrl: string;
3440
+ reasoning: true;
3441
+ input: ("image" | "text")[];
3442
+ cost: {
3443
+ input: number;
3444
+ output: number;
3445
+ cacheRead: number;
3446
+ cacheWrite: number;
3447
+ };
3448
+ contextWindow: number;
3449
+ maxTokens: number;
3450
+ };
3451
+ readonly "gpt-5.2-high": {
3452
+ id: string;
3453
+ name: string;
3454
+ api: "openai-codex-responses";
3455
+ provider: string;
3456
+ baseUrl: string;
3457
+ reasoning: true;
3458
+ input: ("image" | "text")[];
3459
+ cost: {
3460
+ input: number;
3461
+ output: number;
3462
+ cacheRead: number;
3463
+ cacheWrite: number;
3464
+ };
3465
+ contextWindow: number;
3466
+ maxTokens: number;
3467
+ };
3468
+ readonly "gpt-5.2-low": {
3469
+ id: string;
3470
+ name: string;
3471
+ api: "openai-codex-responses";
3472
+ provider: string;
3473
+ baseUrl: string;
3474
+ reasoning: true;
3475
+ input: ("image" | "text")[];
3476
+ cost: {
3477
+ input: number;
3478
+ output: number;
3479
+ cacheRead: number;
3480
+ cacheWrite: number;
3481
+ };
3482
+ contextWindow: number;
3483
+ maxTokens: number;
3484
+ };
3485
+ readonly "gpt-5.2-medium": {
3486
+ id: string;
3487
+ name: string;
3488
+ api: "openai-codex-responses";
3489
+ provider: string;
3490
+ baseUrl: string;
3491
+ reasoning: true;
3492
+ input: ("image" | "text")[];
3493
+ cost: {
3494
+ input: number;
3495
+ output: number;
3496
+ cacheRead: number;
3497
+ cacheWrite: number;
3498
+ };
3499
+ contextWindow: number;
3500
+ maxTokens: number;
3501
+ };
3502
+ readonly "gpt-5.2-none": {
3503
+ id: string;
3504
+ name: string;
3505
+ api: "openai-codex-responses";
3506
+ provider: string;
3507
+ baseUrl: string;
3508
+ reasoning: true;
3509
+ input: ("image" | "text")[];
3510
+ cost: {
3511
+ input: number;
3512
+ output: number;
3513
+ cacheRead: number;
3514
+ cacheWrite: number;
3515
+ };
3516
+ contextWindow: number;
3517
+ maxTokens: number;
3518
+ };
3519
+ readonly "gpt-5.2-xhigh": {
3520
+ id: string;
3521
+ name: string;
3522
+ api: "openai-codex-responses";
3523
+ provider: string;
3524
+ baseUrl: string;
3525
+ reasoning: true;
3526
+ input: ("image" | "text")[];
3527
+ cost: {
3528
+ input: number;
3529
+ output: number;
3530
+ cacheRead: number;
3531
+ cacheWrite: number;
3532
+ };
3533
+ contextWindow: number;
3534
+ maxTokens: number;
3535
+ };
3536
+ };
3537
+ readonly openrouter: {
3538
+ readonly "ai21/jamba-large-1.7": {
2925
3539
  id: string;
2926
3540
  name: string;
2927
3541
  api: "openai-completions";
@@ -2938,13 +3552,13 @@ export declare const MODELS: {
2938
3552
  contextWindow: number;
2939
3553
  maxTokens: number;
2940
3554
  };
2941
- readonly "alibaba/tongyi-deepresearch-30b-a3b": {
3555
+ readonly "ai21/jamba-mini-1.7": {
2942
3556
  id: string;
2943
3557
  name: string;
2944
3558
  api: "openai-completions";
2945
3559
  provider: string;
2946
3560
  baseUrl: string;
2947
- reasoning: true;
3561
+ reasoning: false;
2948
3562
  input: "text"[];
2949
3563
  cost: {
2950
3564
  input: number;
@@ -2955,7 +3569,7 @@ export declare const MODELS: {
2955
3569
  contextWindow: number;
2956
3570
  maxTokens: number;
2957
3571
  };
2958
- readonly "alibaba/tongyi-deepresearch-30b-a3b:free": {
3572
+ readonly "alibaba/tongyi-deepresearch-30b-a3b": {
2959
3573
  id: string;
2960
3574
  name: string;
2961
3575
  api: "openai-completions";
@@ -3941,6 +4555,23 @@ export declare const MODELS: {
3941
4555
  contextWindow: number;
3942
4556
  maxTokens: number;
3943
4557
  };
4558
+ readonly "kwaipilot/kat-coder-pro": {
4559
+ id: string;
4560
+ name: string;
4561
+ api: "openai-completions";
4562
+ provider: string;
4563
+ baseUrl: string;
4564
+ reasoning: false;
4565
+ input: "text"[];
4566
+ cost: {
4567
+ input: number;
4568
+ output: number;
4569
+ cacheRead: number;
4570
+ cacheWrite: number;
4571
+ };
4572
+ contextWindow: number;
4573
+ maxTokens: number;
4574
+ };
3944
4575
  readonly "kwaipilot/kat-coder-pro:free": {
3945
4576
  id: string;
3946
4577
  name: string;
@@ -4859,23 +5490,6 @@ export declare const MODELS: {
4859
5490
  contextWindow: number;
4860
5491
  maxTokens: number;
4861
5492
  };
4862
- readonly "nousresearch/hermes-4-405b": {
4863
- id: string;
4864
- name: string;
4865
- api: "openai-completions";
4866
- provider: string;
4867
- baseUrl: string;
4868
- reasoning: true;
4869
- input: "text"[];
4870
- cost: {
4871
- input: number;
4872
- output: number;
4873
- cacheRead: number;
4874
- cacheWrite: number;
4875
- };
4876
- contextWindow: number;
4877
- maxTokens: number;
4878
- };
4879
5493
  readonly "nousresearch/hermes-4-70b": {
4880
5494
  id: string;
4881
5495
  name: string;
@@ -6593,23 +7207,6 @@ export declare const MODELS: {
6593
7207
  contextWindow: number;
6594
7208
  maxTokens: number;
6595
7209
  };
6596
- readonly "tngtech/tng-r1t-chimera:free": {
6597
- id: string;
6598
- name: string;
6599
- api: "openai-completions";
6600
- provider: string;
6601
- baseUrl: string;
6602
- reasoning: true;
6603
- input: "text"[];
6604
- cost: {
6605
- input: number;
6606
- output: number;
6607
- cacheRead: number;
6608
- cacheWrite: number;
6609
- };
6610
- contextWindow: number;
6611
- maxTokens: number;
6612
- };
6613
7210
  readonly "x-ai/grok-3": {
6614
7211
  id: string;
6615
7212
  name: string;