@mariozechner/pi-ai 0.67.68 → 0.68.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (43) hide show
  1. package/README.md +3 -1
  2. package/dist/models.generated.d.ts +64 -35
  3. package/dist/models.generated.d.ts.map +1 -1
  4. package/dist/models.generated.js +77 -56
  5. package/dist/models.generated.js.map +1 -1
  6. package/dist/providers/amazon-bedrock.d.ts.map +1 -1
  7. package/dist/providers/amazon-bedrock.js +9 -1
  8. package/dist/providers/amazon-bedrock.js.map +1 -1
  9. package/dist/providers/anthropic.d.ts.map +1 -1
  10. package/dist/providers/anthropic.js +1 -2
  11. package/dist/providers/anthropic.js.map +1 -1
  12. package/dist/providers/google-shared.d.ts.map +1 -1
  13. package/dist/providers/google-shared.js +30 -4
  14. package/dist/providers/google-shared.js.map +1 -1
  15. package/dist/providers/openai-completions.d.ts +5 -1
  16. package/dist/providers/openai-completions.d.ts.map +1 -1
  17. package/dist/providers/openai-completions.js +149 -60
  18. package/dist/providers/openai-completions.js.map +1 -1
  19. package/dist/providers/openai-responses-shared.d.ts.map +1 -1
  20. package/dist/providers/openai-responses-shared.js +2 -5
  21. package/dist/providers/openai-responses-shared.js.map +1 -1
  22. package/dist/providers/simple-options.d.ts.map +1 -1
  23. package/dist/providers/simple-options.js +1 -1
  24. package/dist/providers/simple-options.js.map +1 -1
  25. package/dist/providers/transform-messages.d.ts.map +1 -1
  26. package/dist/providers/transform-messages.js +41 -2
  27. package/dist/providers/transform-messages.js.map +1 -1
  28. package/dist/types.d.ts +4 -0
  29. package/dist/types.d.ts.map +1 -1
  30. package/dist/types.js.map +1 -1
  31. package/dist/utils/oauth/anthropic.d.ts.map +1 -1
  32. package/dist/utils/oauth/anthropic.js +1 -1
  33. package/dist/utils/oauth/anthropic.js.map +1 -1
  34. package/dist/utils/oauth/google-antigravity.d.ts.map +1 -1
  35. package/dist/utils/oauth/google-antigravity.js +2 -1
  36. package/dist/utils/oauth/google-antigravity.js.map +1 -1
  37. package/dist/utils/oauth/google-gemini-cli.d.ts.map +1 -1
  38. package/dist/utils/oauth/google-gemini-cli.js +2 -1
  39. package/dist/utils/oauth/google-gemini-cli.js.map +1 -1
  40. package/dist/utils/oauth/openai-codex.d.ts.map +1 -1
  41. package/dist/utils/oauth/openai-codex.js +3 -2
  42. package/dist/utils/oauth/openai-codex.js.map +1 -1
  43. package/package.json +1 -1
package/README.md CHANGED
@@ -851,11 +851,13 @@ interface OpenAICompletionsCompat {
851
851
  supportsReasoningEffort?: boolean; // Whether provider supports `reasoning_effort` (default: true)
852
852
  supportsUsageInStreaming?: boolean; // Whether provider supports `stream_options: { include_usage: true }` (default: true)
853
853
  supportsStrictMode?: boolean; // Whether provider supports `strict` in tool definitions (default: true)
854
+ sendSessionAffinityHeaders?: boolean; // Whether to send `session_id`, `x-client-request-id`, and `x-session-affinity` from `sessionId` when caching is enabled (default: false)
854
855
  maxTokensField?: 'max_completion_tokens' | 'max_tokens'; // Which field name to use (default: max_completion_tokens)
855
856
  requiresToolResultName?: boolean; // Whether tool results require the `name` field (default: false)
856
857
  requiresAssistantAfterToolResult?: boolean; // Whether tool results must be followed by an assistant message (default: false)
857
858
  requiresThinkingAsText?: boolean; // Whether thinking blocks must be converted to text (default: false)
858
- thinkingFormat?: 'openai' | 'zai' | 'qwen'; // Format for reasoning param: 'openai' uses reasoning_effort, 'zai' uses thinking: { type: "enabled" }, 'qwen' uses enable_thinking: boolean (default: openai)
859
+ thinkingFormat?: 'openai' | 'zai' | 'qwen' | 'qwen-chat-template'; // Format for reasoning param: 'openai' uses reasoning_effort, 'zai' uses thinking: { type: "enabled" }, 'qwen' uses enable_thinking: boolean, 'qwen-chat-template' uses chat_template_kwargs.enable_thinking (default: openai)
860
+ cacheControlFormat?: 'anthropic'; // Anthropic-style cache_control on system prompt, last tool, and last user/assistant text content
859
861
  openRouterRouting?: OpenRouterRouting; // OpenRouter routing preferences (default: {})
860
862
  vercelGatewayRouting?: VercelGatewayRouting; // Vercel AI Gateway routing preferences (default: {})
861
863
  }
@@ -6898,6 +6898,9 @@ export declare const MODELS: {
6898
6898
  api: "openai-completions";
6899
6899
  provider: string;
6900
6900
  baseUrl: string;
6901
+ compat: {
6902
+ cacheControlFormat: "anthropic";
6903
+ };
6901
6904
  reasoning: true;
6902
6905
  input: ("image" | "text")[];
6903
6906
  cost: {
@@ -6915,6 +6918,9 @@ export declare const MODELS: {
6915
6918
  api: "openai-completions";
6916
6919
  provider: string;
6917
6920
  baseUrl: string;
6921
+ compat: {
6922
+ cacheControlFormat: "anthropic";
6923
+ };
6918
6924
  reasoning: true;
6919
6925
  input: ("image" | "text")[];
6920
6926
  cost: {
@@ -6979,6 +6985,23 @@ export declare const MODELS: {
6979
6985
  contextWindow: number;
6980
6986
  maxTokens: number;
6981
6987
  };
6988
+ readonly "kimi-k2.6": {
6989
+ id: string;
6990
+ name: string;
6991
+ api: "openai-completions";
6992
+ provider: string;
6993
+ baseUrl: string;
6994
+ reasoning: true;
6995
+ input: ("image" | "text")[];
6996
+ cost: {
6997
+ input: number;
6998
+ output: number;
6999
+ cacheRead: number;
7000
+ cacheWrite: number;
7001
+ };
7002
+ contextWindow: number;
7003
+ maxTokens: number;
7004
+ };
6982
7005
  readonly "mimo-v2-omni": {
6983
7006
  id: string;
6984
7007
  name: string;
@@ -7053,6 +7076,9 @@ export declare const MODELS: {
7053
7076
  api: "openai-completions";
7054
7077
  provider: string;
7055
7078
  baseUrl: string;
7079
+ compat: {
7080
+ cacheControlFormat: "anthropic";
7081
+ };
7056
7082
  reasoning: true;
7057
7083
  input: ("image" | "text")[];
7058
7084
  cost: {
@@ -7070,6 +7096,9 @@ export declare const MODELS: {
7070
7096
  api: "openai-completions";
7071
7097
  provider: string;
7072
7098
  baseUrl: string;
7099
+ compat: {
7100
+ cacheControlFormat: "anthropic";
7101
+ };
7073
7102
  reasoning: true;
7074
7103
  input: ("image" | "text")[];
7075
7104
  cost: {
@@ -8222,23 +8251,6 @@ export declare const MODELS: {
8222
8251
  contextWindow: number;
8223
8252
  maxTokens: number;
8224
8253
  };
8225
- readonly "meta-llama/llama-4-maverick": {
8226
- id: string;
8227
- name: string;
8228
- api: "openai-completions";
8229
- provider: string;
8230
- baseUrl: string;
8231
- reasoning: false;
8232
- input: ("image" | "text")[];
8233
- cost: {
8234
- input: number;
8235
- output: number;
8236
- cacheRead: number;
8237
- cacheWrite: number;
8238
- };
8239
- contextWindow: number;
8240
- maxTokens: number;
8241
- };
8242
8254
  readonly "meta-llama/llama-4-scout": {
8243
8255
  id: string;
8244
8256
  name: string;
@@ -8800,6 +8812,23 @@ export declare const MODELS: {
8800
8812
  contextWindow: number;
8801
8813
  maxTokens: number;
8802
8814
  };
8815
+ readonly "moonshotai/kimi-k2.6": {
8816
+ id: string;
8817
+ name: string;
8818
+ api: "openai-completions";
8819
+ provider: string;
8820
+ baseUrl: string;
8821
+ reasoning: true;
8822
+ input: ("image" | "text")[];
8823
+ cost: {
8824
+ input: number;
8825
+ output: number;
8826
+ cacheRead: number;
8827
+ cacheWrite: number;
8828
+ };
8829
+ contextWindow: number;
8830
+ maxTokens: number;
8831
+ };
8803
8832
  readonly "nex-agi/deepseek-v3.1-nex-n1": {
8804
8833
  id: string;
8805
8834
  name: string;
@@ -9276,23 +9305,6 @@ export declare const MODELS: {
9276
9305
  contextWindow: number;
9277
9306
  maxTokens: number;
9278
9307
  };
9279
- readonly "openai/gpt-4o:extended": {
9280
- id: string;
9281
- name: string;
9282
- api: "openai-completions";
9283
- provider: string;
9284
- baseUrl: string;
9285
- reasoning: false;
9286
- input: ("image" | "text")[];
9287
- cost: {
9288
- input: number;
9289
- output: number;
9290
- cacheRead: number;
9291
- cacheWrite: number;
9292
- };
9293
- contextWindow: number;
9294
- maxTokens: number;
9295
- };
9296
9308
  readonly "openai/gpt-5": {
9297
9309
  id: string;
9298
9310
  name: string;
@@ -10183,7 +10195,7 @@ export declare const MODELS: {
10183
10195
  api: "openai-completions";
10184
10196
  provider: string;
10185
10197
  baseUrl: string;
10186
- reasoning: true;
10198
+ reasoning: false;
10187
10199
  input: "text"[];
10188
10200
  cost: {
10189
10201
  input: number;
@@ -12746,6 +12758,23 @@ export declare const MODELS: {
12746
12758
  contextWindow: number;
12747
12759
  maxTokens: number;
12748
12760
  };
12761
+ readonly "moonshotai/kimi-k2.6": {
12762
+ id: string;
12763
+ name: string;
12764
+ api: "anthropic-messages";
12765
+ provider: string;
12766
+ baseUrl: string;
12767
+ reasoning: true;
12768
+ input: ("image" | "text")[];
12769
+ cost: {
12770
+ input: number;
12771
+ output: number;
12772
+ cacheRead: number;
12773
+ cacheWrite: number;
12774
+ };
12775
+ contextWindow: number;
12776
+ maxTokens: number;
12777
+ };
12749
12778
  readonly "nvidia/nemotron-nano-12b-v2-vl": {
12750
12779
  id: string;
12751
12780
  name: string;