@hebo-ai/gateway 0.10.7 → 0.11.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (111) hide show
  1. package/README.md +97 -84
  2. package/dist/config.js +21 -4
  3. package/dist/endpoints/chat-completions/handler.js +9 -5
  4. package/dist/endpoints/conversations/handler.js +3 -3
  5. package/dist/endpoints/embeddings/handler.js +2 -2
  6. package/dist/endpoints/messages/converters.js +17 -21
  7. package/dist/endpoints/messages/handler.js +5 -5
  8. package/dist/endpoints/responses/handler.js +9 -5
  9. package/dist/index.d.ts +1 -0
  10. package/dist/index.js +1 -0
  11. package/dist/models/alibaba/index.d.ts +2 -0
  12. package/dist/models/alibaba/index.js +2 -0
  13. package/dist/models/alibaba/middleware.d.ts +2 -0
  14. package/dist/models/alibaba/middleware.js +31 -0
  15. package/dist/models/alibaba/presets.d.ts +900 -0
  16. package/dist/models/alibaba/presets.js +252 -0
  17. package/dist/models/amazon/presets.d.ts +31 -31
  18. package/dist/models/anthropic/presets.d.ts +68 -68
  19. package/dist/models/cohere/presets.d.ts +38 -38
  20. package/dist/models/deepseek/index.d.ts +2 -0
  21. package/dist/models/deepseek/index.js +2 -0
  22. package/dist/models/deepseek/middleware.d.ts +2 -0
  23. package/dist/models/deepseek/middleware.js +25 -0
  24. package/dist/models/deepseek/presets.d.ts +51 -0
  25. package/dist/models/deepseek/presets.js +33 -0
  26. package/dist/models/google/presets.d.ts +94 -94
  27. package/dist/models/google/presets.js +3 -1
  28. package/dist/models/meta/presets.d.ts +84 -84
  29. package/dist/models/meta/presets.js +11 -12
  30. package/dist/models/minimax/index.d.ts +2 -0
  31. package/dist/models/minimax/index.js +2 -0
  32. package/dist/models/minimax/middleware.d.ts +2 -0
  33. package/dist/models/minimax/middleware.js +43 -0
  34. package/dist/models/minimax/presets.d.ts +99 -0
  35. package/dist/models/minimax/presets.js +47 -0
  36. package/dist/models/moonshot/index.d.ts +2 -0
  37. package/dist/models/moonshot/index.js +2 -0
  38. package/dist/models/moonshot/middleware.d.ts +2 -0
  39. package/dist/models/moonshot/middleware.js +32 -0
  40. package/dist/models/moonshot/presets.d.ts +104 -0
  41. package/dist/models/moonshot/presets.js +36 -0
  42. package/dist/models/openai/presets.d.ts +128 -128
  43. package/dist/models/openai/presets.js +1 -1
  44. package/dist/models/types.d.ts +1 -1
  45. package/dist/models/types.js +39 -0
  46. package/dist/models/voyage/presets.d.ts +46 -46
  47. package/dist/models/xai/index.d.ts +2 -0
  48. package/dist/models/xai/index.js +2 -0
  49. package/dist/models/xai/middleware.d.ts +2 -0
  50. package/dist/models/xai/middleware.js +37 -0
  51. package/dist/models/xai/presets.d.ts +210 -0
  52. package/dist/models/xai/presets.js +55 -0
  53. package/dist/models/zai/index.d.ts +2 -0
  54. package/dist/models/zai/index.js +2 -0
  55. package/dist/models/zai/middleware.d.ts +2 -0
  56. package/dist/models/zai/middleware.js +25 -0
  57. package/dist/models/zai/presets.d.ts +141 -0
  58. package/dist/models/zai/presets.js +41 -0
  59. package/dist/providers/alibaba/canonical.d.ts +3 -0
  60. package/dist/providers/alibaba/canonical.js +13 -0
  61. package/dist/providers/alibaba/index.d.ts +1 -0
  62. package/dist/providers/alibaba/index.js +1 -0
  63. package/dist/providers/bedrock/canonical.js +3 -0
  64. package/dist/providers/chutes/canonical.d.ts +3 -0
  65. package/dist/providers/chutes/canonical.js +14 -0
  66. package/dist/providers/chutes/index.d.ts +1 -0
  67. package/dist/providers/chutes/index.js +1 -0
  68. package/dist/providers/deepinfra/canonical.d.ts +3 -0
  69. package/dist/providers/deepinfra/canonical.js +40 -0
  70. package/dist/providers/deepinfra/index.d.ts +1 -0
  71. package/dist/providers/deepinfra/index.js +1 -0
  72. package/dist/providers/deepseek/canonical.d.ts +3 -0
  73. package/dist/providers/deepseek/canonical.js +9 -0
  74. package/dist/providers/deepseek/index.d.ts +1 -0
  75. package/dist/providers/deepseek/index.js +1 -0
  76. package/dist/providers/fireworks/canonical.d.ts +3 -0
  77. package/dist/providers/fireworks/canonical.js +27 -0
  78. package/dist/providers/fireworks/index.d.ts +2 -0
  79. package/dist/providers/fireworks/index.js +2 -0
  80. package/dist/providers/fireworks/middleware.d.ts +2 -0
  81. package/dist/providers/fireworks/middleware.js +35 -0
  82. package/dist/providers/groq/canonical.js +1 -1
  83. package/dist/providers/minimax/canonical.d.ts +3 -0
  84. package/dist/providers/minimax/canonical.js +9 -0
  85. package/dist/providers/minimax/index.d.ts +1 -0
  86. package/dist/providers/minimax/index.js +1 -0
  87. package/dist/providers/moonshot/canonical.d.ts +3 -0
  88. package/dist/providers/moonshot/canonical.js +6 -0
  89. package/dist/providers/moonshot/index.d.ts +1 -0
  90. package/dist/providers/moonshot/index.js +1 -0
  91. package/dist/providers/togetherai/canonical.d.ts +3 -0
  92. package/dist/providers/togetherai/canonical.js +24 -0
  93. package/dist/providers/togetherai/index.d.ts +1 -0
  94. package/dist/providers/togetherai/index.js +1 -0
  95. package/dist/providers/types.d.ts +1 -1
  96. package/dist/providers/types.js +10 -0
  97. package/dist/providers/vertex/canonical.js +5 -1
  98. package/dist/providers/xai/canonical.d.ts +3 -0
  99. package/dist/providers/xai/canonical.js +12 -0
  100. package/dist/providers/xai/index.d.ts +1 -0
  101. package/dist/providers/xai/index.js +1 -0
  102. package/dist/providers/zai/canonical.d.ts +3 -0
  103. package/dist/providers/zai/canonical.js +10 -0
  104. package/dist/providers/zai/index.d.ts +1 -0
  105. package/dist/providers/zai/index.js +1 -0
  106. package/dist/types.d.ts +30 -16
  107. package/dist/utils/index.d.ts +8 -0
  108. package/dist/utils/index.js +8 -0
  109. package/dist/utils/request.d.ts +2 -1
  110. package/dist/utils/request.js +35 -7
  111. package/package.json +133 -9
@@ -80,7 +80,7 @@ const GEMMA3_BASE = {
80
80
  capabilities: ["tool_call", "structured_output", "temperature"],
81
81
  context: 131072,
82
82
  knowledge: "2025-01",
83
- providers: ["vertex", "bedrock"],
83
+ providers: ["vertex", "bedrock", "deepinfra"],
84
84
  };
85
85
  const GEMMA4_BASE = {
86
86
  modalities: {
@@ -141,12 +141,14 @@ export const gemma426bA4b = presetFor()("google/gemma-4-26b-a4b", {
141
141
  name: "Gemma 4 26B-A4B",
142
142
  created: "2026-04-02",
143
143
  context: 262144,
144
+ providers: ["vertex", "deepinfra"],
144
145
  });
145
146
  export const gemma431b = presetFor()("google/gemma-4-31b", {
146
147
  ...GEMMA4_BASE,
147
148
  name: "Gemma 4 31B",
148
149
  created: "2026-04-02",
149
150
  context: 262144,
151
+ providers: ["vertex", "deepinfra", "togetherai"],
150
152
  });
151
153
  const gemmaAtomic = {
152
154
  v3: [gemma31b, gemma34b, gemma312b, gemma327b],
@@ -1,41 +1,41 @@
1
1
  import type { CatalogModel } from "../types";
2
- export declare const llama31_8b: import("../../utils/preset").Preset<"meta/llama-3.1-8b", CatalogModel, {
2
+ export declare const llama31_8b: import("../../utils").Preset<"meta/llama-3.1-8b", CatalogModel, {
3
3
  name: string;
4
4
  created: string;
5
5
  knowledge: string;
6
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
6
7
  modalities: {
7
8
  input: readonly ["text", "file"];
8
9
  output: readonly ["text"];
9
10
  };
10
11
  capabilities: readonly ["attachments", "tool_call", "temperature"];
11
12
  context: number;
12
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
13
13
  }>;
14
- export declare const llama31_70b: import("../../utils/preset").Preset<"meta/llama-3.1-70b", CatalogModel, {
14
+ export declare const llama31_70b: import("../../utils").Preset<"meta/llama-3.1-70b", CatalogModel, {
15
15
  name: string;
16
16
  created: string;
17
17
  knowledge: string;
18
+ providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
18
19
  modalities: {
19
20
  input: readonly ["text", "file"];
20
21
  output: readonly ["text"];
21
22
  };
22
23
  capabilities: readonly ["attachments", "tool_call", "temperature"];
23
24
  context: number;
24
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
25
25
  }>;
26
- export declare const llama31_405b: import("../../utils/preset").Preset<"meta/llama-3.1-405b", CatalogModel, {
26
+ export declare const llama31_405b: import("../../utils").Preset<"meta/llama-3.1-405b", CatalogModel, {
27
27
  name: string;
28
28
  created: string;
29
29
  knowledge: string;
30
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "togetherai" | "vertex")[];
30
31
  modalities: {
31
32
  input: readonly ["text", "file"];
32
33
  output: readonly ["text"];
33
34
  };
34
35
  capabilities: readonly ["attachments", "tool_call", "temperature"];
35
36
  context: number;
36
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
37
37
  }>;
38
- export declare const llama32_1b: import("../../utils/preset").Preset<"meta/llama-3.2-1b", CatalogModel, {
38
+ export declare const llama32_1b: import("../../utils").Preset<"meta/llama-3.2-1b", CatalogModel, {
39
39
  name: string;
40
40
  created: string;
41
41
  knowledge: string;
@@ -45,138 +45,138 @@ export declare const llama32_1b: import("../../utils/preset").Preset<"meta/llama
45
45
  };
46
46
  capabilities: readonly ["attachments", "tool_call", "temperature"];
47
47
  context: number;
48
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
48
+ providers: readonly ["bedrock", "vertex", "azure"];
49
49
  }>;
50
- export declare const llama32_3b: import("../../utils/preset").Preset<"meta/llama-3.2-3b", CatalogModel, {
50
+ export declare const llama32_3b: import("../../utils").Preset<"meta/llama-3.2-3b", CatalogModel, {
51
51
  name: string;
52
52
  created: string;
53
53
  knowledge: string;
54
+ providers: ("azure" | "bedrock" | "fireworks" | "togetherai" | "vertex")[];
54
55
  modalities: {
55
56
  input: readonly ["text", "file"];
56
57
  output: readonly ["text"];
57
58
  };
58
59
  capabilities: readonly ["attachments", "tool_call", "temperature"];
59
60
  context: number;
60
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
61
61
  }>;
62
- export declare const llama32_11b: import("../../utils/preset").Preset<"meta/llama-3.2-11b", CatalogModel, {
62
+ export declare const llama32_11b: import("../../utils").Preset<"meta/llama-3.2-11b", CatalogModel, {
63
63
  name: string;
64
64
  created: string;
65
65
  knowledge: string;
66
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "vertex")[];
66
67
  modalities: {
67
68
  input: readonly ["text", "file"];
68
69
  output: readonly ["text"];
69
70
  };
70
71
  capabilities: readonly ["attachments", "tool_call", "temperature"];
71
72
  context: number;
72
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
73
73
  }>;
74
- export declare const llama32_90b: import("../../utils/preset").Preset<"meta/llama-3.2-90b", CatalogModel, {
74
+ export declare const llama32_90b: import("../../utils").Preset<"meta/llama-3.2-90b", CatalogModel, {
75
75
  name: string;
76
76
  created: string;
77
77
  knowledge: string;
78
+ providers: ("azure" | "bedrock" | "deepinfra" | "vertex")[];
78
79
  modalities: {
79
80
  input: readonly ["text", "file"];
80
81
  output: readonly ["text"];
81
82
  };
82
83
  capabilities: readonly ["attachments", "tool_call", "temperature"];
83
84
  context: number;
84
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
85
85
  }>;
86
- export declare const llama33_70b: import("../../utils/preset").Preset<"meta/llama-3.3-70b", CatalogModel, {
86
+ export declare const llama33_70b: import("../../utils").Preset<"meta/llama-3.3-70b", CatalogModel, {
87
87
  name: string;
88
88
  created: string;
89
89
  knowledge: string;
90
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
90
91
  modalities: {
91
92
  input: readonly ["text", "file"];
92
93
  output: readonly ["text"];
93
94
  };
94
95
  capabilities: readonly ["attachments", "tool_call", "temperature"];
95
96
  context: number;
96
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
97
97
  }>;
98
- export declare const llama4Scout: import("../../utils/preset").Preset<"meta/llama-4-scout", CatalogModel, {
98
+ export declare const llama4Scout: import("../../utils").Preset<"meta/llama-4-scout", CatalogModel, {
99
99
  name: string;
100
100
  created: string;
101
101
  knowledge: string;
102
+ providers: ("azure" | "bedrock" | "deepinfra" | "groq" | "vertex")[];
102
103
  modalities: {
103
104
  input: readonly ["text", "image", "file"];
104
105
  output: readonly ["text"];
105
106
  };
106
107
  capabilities: readonly ["attachments", "tool_call", "temperature"];
107
108
  context: number;
108
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
109
109
  }>;
110
- export declare const llama4Maverick: import("../../utils/preset").Preset<"meta/llama-4-maverick", CatalogModel, {
110
+ export declare const llama4Maverick: import("../../utils").Preset<"meta/llama-4-maverick", CatalogModel, {
111
111
  name: string;
112
112
  created: string;
113
113
  knowledge: string;
114
+ providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
114
115
  modalities: {
115
116
  input: readonly ["text", "image", "file"];
116
117
  output: readonly ["text"];
117
118
  };
118
119
  capabilities: readonly ["attachments", "tool_call", "temperature"];
119
120
  context: number;
120
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
121
121
  }>;
122
122
  export declare const llama: {
123
- readonly latest: readonly [import("../../utils/preset").Preset<"meta/llama-4-scout", CatalogModel, {
123
+ readonly latest: readonly [import("../../utils").Preset<"meta/llama-4-scout", CatalogModel, {
124
124
  name: string;
125
125
  created: string;
126
126
  knowledge: string;
127
+ providers: ("azure" | "bedrock" | "deepinfra" | "groq" | "vertex")[];
127
128
  modalities: {
128
129
  input: readonly ["text", "image", "file"];
129
130
  output: readonly ["text"];
130
131
  };
131
132
  capabilities: readonly ["attachments", "tool_call", "temperature"];
132
133
  context: number;
133
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
134
- }>, import("../../utils/preset").Preset<"meta/llama-4-maverick", CatalogModel, {
134
+ }>, import("../../utils").Preset<"meta/llama-4-maverick", CatalogModel, {
135
135
  name: string;
136
136
  created: string;
137
137
  knowledge: string;
138
+ providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
138
139
  modalities: {
139
140
  input: readonly ["text", "image", "file"];
140
141
  output: readonly ["text"];
141
142
  };
142
143
  capabilities: readonly ["attachments", "tool_call", "temperature"];
143
144
  context: number;
144
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
145
145
  }>];
146
- readonly all: (import("../../utils/preset").Preset<"meta/llama-3.1-8b", CatalogModel, {
146
+ readonly all: (import("../../utils").Preset<"meta/llama-3.1-8b", CatalogModel, {
147
147
  name: string;
148
148
  created: string;
149
149
  knowledge: string;
150
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
150
151
  modalities: {
151
152
  input: readonly ["text", "file"];
152
153
  output: readonly ["text"];
153
154
  };
154
155
  capabilities: readonly ["attachments", "tool_call", "temperature"];
155
156
  context: number;
156
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
157
- }> | import("../../utils/preset").Preset<"meta/llama-3.1-70b", CatalogModel, {
157
+ }> | import("../../utils").Preset<"meta/llama-3.1-70b", CatalogModel, {
158
158
  name: string;
159
159
  created: string;
160
160
  knowledge: string;
161
+ providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
161
162
  modalities: {
162
163
  input: readonly ["text", "file"];
163
164
  output: readonly ["text"];
164
165
  };
165
166
  capabilities: readonly ["attachments", "tool_call", "temperature"];
166
167
  context: number;
167
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
168
- }> | import("../../utils/preset").Preset<"meta/llama-3.1-405b", CatalogModel, {
168
+ }> | import("../../utils").Preset<"meta/llama-3.1-405b", CatalogModel, {
169
169
  name: string;
170
170
  created: string;
171
171
  knowledge: string;
172
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "togetherai" | "vertex")[];
172
173
  modalities: {
173
174
  input: readonly ["text", "file"];
174
175
  output: readonly ["text"];
175
176
  };
176
177
  capabilities: readonly ["attachments", "tool_call", "temperature"];
177
178
  context: number;
178
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
179
- }> | import("../../utils/preset").Preset<"meta/llama-3.2-1b", CatalogModel, {
179
+ }> | import("../../utils").Preset<"meta/llama-3.2-1b", CatalogModel, {
180
180
  name: string;
181
181
  created: string;
182
182
  knowledge: string;
@@ -186,108 +186,108 @@ export declare const llama: {
186
186
  };
187
187
  capabilities: readonly ["attachments", "tool_call", "temperature"];
188
188
  context: number;
189
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
190
- }> | import("../../utils/preset").Preset<"meta/llama-3.2-3b", CatalogModel, {
189
+ providers: readonly ["bedrock", "vertex", "azure"];
190
+ }> | import("../../utils").Preset<"meta/llama-3.2-3b", CatalogModel, {
191
191
  name: string;
192
192
  created: string;
193
193
  knowledge: string;
194
+ providers: ("azure" | "bedrock" | "fireworks" | "togetherai" | "vertex")[];
194
195
  modalities: {
195
196
  input: readonly ["text", "file"];
196
197
  output: readonly ["text"];
197
198
  };
198
199
  capabilities: readonly ["attachments", "tool_call", "temperature"];
199
200
  context: number;
200
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
201
- }> | import("../../utils/preset").Preset<"meta/llama-3.2-11b", CatalogModel, {
201
+ }> | import("../../utils").Preset<"meta/llama-3.2-11b", CatalogModel, {
202
202
  name: string;
203
203
  created: string;
204
204
  knowledge: string;
205
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "vertex")[];
205
206
  modalities: {
206
207
  input: readonly ["text", "file"];
207
208
  output: readonly ["text"];
208
209
  };
209
210
  capabilities: readonly ["attachments", "tool_call", "temperature"];
210
211
  context: number;
211
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
212
- }> | import("../../utils/preset").Preset<"meta/llama-3.2-90b", CatalogModel, {
212
+ }> | import("../../utils").Preset<"meta/llama-3.2-90b", CatalogModel, {
213
213
  name: string;
214
214
  created: string;
215
215
  knowledge: string;
216
+ providers: ("azure" | "bedrock" | "deepinfra" | "vertex")[];
216
217
  modalities: {
217
218
  input: readonly ["text", "file"];
218
219
  output: readonly ["text"];
219
220
  };
220
221
  capabilities: readonly ["attachments", "tool_call", "temperature"];
221
222
  context: number;
222
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
223
- }> | import("../../utils/preset").Preset<"meta/llama-3.3-70b", CatalogModel, {
223
+ }> | import("../../utils").Preset<"meta/llama-3.3-70b", CatalogModel, {
224
224
  name: string;
225
225
  created: string;
226
226
  knowledge: string;
227
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
227
228
  modalities: {
228
229
  input: readonly ["text", "file"];
229
230
  output: readonly ["text"];
230
231
  };
231
232
  capabilities: readonly ["attachments", "tool_call", "temperature"];
232
233
  context: number;
233
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
234
- }> | import("../../utils/preset").Preset<"meta/llama-4-scout", CatalogModel, {
234
+ }> | import("../../utils").Preset<"meta/llama-4-scout", CatalogModel, {
235
235
  name: string;
236
236
  created: string;
237
237
  knowledge: string;
238
+ providers: ("azure" | "bedrock" | "deepinfra" | "groq" | "vertex")[];
238
239
  modalities: {
239
240
  input: readonly ["text", "image", "file"];
240
241
  output: readonly ["text"];
241
242
  };
242
243
  capabilities: readonly ["attachments", "tool_call", "temperature"];
243
244
  context: number;
244
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
245
- }> | import("../../utils/preset").Preset<"meta/llama-4-maverick", CatalogModel, {
245
+ }> | import("../../utils").Preset<"meta/llama-4-maverick", CatalogModel, {
246
246
  name: string;
247
247
  created: string;
248
248
  knowledge: string;
249
+ providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
249
250
  modalities: {
250
251
  input: readonly ["text", "image", "file"];
251
252
  output: readonly ["text"];
252
253
  };
253
254
  capabilities: readonly ["attachments", "tool_call", "temperature"];
254
255
  context: number;
255
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
256
256
  }>)[];
257
- readonly "v3.x": readonly [import("../../utils/preset").Preset<"meta/llama-3.1-8b", CatalogModel, {
257
+ readonly "v3.x": readonly [import("../../utils").Preset<"meta/llama-3.1-8b", CatalogModel, {
258
258
  name: string;
259
259
  created: string;
260
260
  knowledge: string;
261
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
261
262
  modalities: {
262
263
  input: readonly ["text", "file"];
263
264
  output: readonly ["text"];
264
265
  };
265
266
  capabilities: readonly ["attachments", "tool_call", "temperature"];
266
267
  context: number;
267
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
268
- }>, import("../../utils/preset").Preset<"meta/llama-3.1-70b", CatalogModel, {
268
+ }>, import("../../utils").Preset<"meta/llama-3.1-70b", CatalogModel, {
269
269
  name: string;
270
270
  created: string;
271
271
  knowledge: string;
272
+ providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
272
273
  modalities: {
273
274
  input: readonly ["text", "file"];
274
275
  output: readonly ["text"];
275
276
  };
276
277
  capabilities: readonly ["attachments", "tool_call", "temperature"];
277
278
  context: number;
278
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
279
- }>, import("../../utils/preset").Preset<"meta/llama-3.1-405b", CatalogModel, {
279
+ }>, import("../../utils").Preset<"meta/llama-3.1-405b", CatalogModel, {
280
280
  name: string;
281
281
  created: string;
282
282
  knowledge: string;
283
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "togetherai" | "vertex")[];
283
284
  modalities: {
284
285
  input: readonly ["text", "file"];
285
286
  output: readonly ["text"];
286
287
  };
287
288
  capabilities: readonly ["attachments", "tool_call", "temperature"];
288
289
  context: number;
289
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
290
- }>, import("../../utils/preset").Preset<"meta/llama-3.2-1b", CatalogModel, {
290
+ }>, import("../../utils").Preset<"meta/llama-3.2-1b", CatalogModel, {
291
291
  name: string;
292
292
  created: string;
293
293
  knowledge: string;
@@ -297,110 +297,110 @@ export declare const llama: {
297
297
  };
298
298
  capabilities: readonly ["attachments", "tool_call", "temperature"];
299
299
  context: number;
300
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
301
- }>, import("../../utils/preset").Preset<"meta/llama-3.2-3b", CatalogModel, {
300
+ providers: readonly ["bedrock", "vertex", "azure"];
301
+ }>, import("../../utils").Preset<"meta/llama-3.2-3b", CatalogModel, {
302
302
  name: string;
303
303
  created: string;
304
304
  knowledge: string;
305
+ providers: ("azure" | "bedrock" | "fireworks" | "togetherai" | "vertex")[];
305
306
  modalities: {
306
307
  input: readonly ["text", "file"];
307
308
  output: readonly ["text"];
308
309
  };
309
310
  capabilities: readonly ["attachments", "tool_call", "temperature"];
310
311
  context: number;
311
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
312
- }>, import("../../utils/preset").Preset<"meta/llama-3.2-11b", CatalogModel, {
312
+ }>, import("../../utils").Preset<"meta/llama-3.2-11b", CatalogModel, {
313
313
  name: string;
314
314
  created: string;
315
315
  knowledge: string;
316
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "vertex")[];
316
317
  modalities: {
317
318
  input: readonly ["text", "file"];
318
319
  output: readonly ["text"];
319
320
  };
320
321
  capabilities: readonly ["attachments", "tool_call", "temperature"];
321
322
  context: number;
322
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
323
- }>, import("../../utils/preset").Preset<"meta/llama-3.2-90b", CatalogModel, {
323
+ }>, import("../../utils").Preset<"meta/llama-3.2-90b", CatalogModel, {
324
324
  name: string;
325
325
  created: string;
326
326
  knowledge: string;
327
+ providers: ("azure" | "bedrock" | "deepinfra" | "vertex")[];
327
328
  modalities: {
328
329
  input: readonly ["text", "file"];
329
330
  output: readonly ["text"];
330
331
  };
331
332
  capabilities: readonly ["attachments", "tool_call", "temperature"];
332
333
  context: number;
333
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
334
- }>, import("../../utils/preset").Preset<"meta/llama-3.3-70b", CatalogModel, {
334
+ }>, import("../../utils").Preset<"meta/llama-3.3-70b", CatalogModel, {
335
335
  name: string;
336
336
  created: string;
337
337
  knowledge: string;
338
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
338
339
  modalities: {
339
340
  input: readonly ["text", "file"];
340
341
  output: readonly ["text"];
341
342
  };
342
343
  capabilities: readonly ["attachments", "tool_call", "temperature"];
343
344
  context: number;
344
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
345
345
  }>];
346
- readonly "v4.x": readonly [import("../../utils/preset").Preset<"meta/llama-4-scout", CatalogModel, {
346
+ readonly "v4.x": readonly [import("../../utils").Preset<"meta/llama-4-scout", CatalogModel, {
347
347
  name: string;
348
348
  created: string;
349
349
  knowledge: string;
350
+ providers: ("azure" | "bedrock" | "deepinfra" | "groq" | "vertex")[];
350
351
  modalities: {
351
352
  input: readonly ["text", "image", "file"];
352
353
  output: readonly ["text"];
353
354
  };
354
355
  capabilities: readonly ["attachments", "tool_call", "temperature"];
355
356
  context: number;
356
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
357
- }>, import("../../utils/preset").Preset<"meta/llama-4-maverick", CatalogModel, {
357
+ }>, import("../../utils").Preset<"meta/llama-4-maverick", CatalogModel, {
358
358
  name: string;
359
359
  created: string;
360
360
  knowledge: string;
361
+ providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
361
362
  modalities: {
362
363
  input: readonly ["text", "image", "file"];
363
364
  output: readonly ["text"];
364
365
  };
365
366
  capabilities: readonly ["attachments", "tool_call", "temperature"];
366
367
  context: number;
367
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
368
368
  }>];
369
- readonly "v3.1": readonly [import("../../utils/preset").Preset<"meta/llama-3.1-8b", CatalogModel, {
369
+ readonly "v3.1": readonly [import("../../utils").Preset<"meta/llama-3.1-8b", CatalogModel, {
370
370
  name: string;
371
371
  created: string;
372
372
  knowledge: string;
373
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
373
374
  modalities: {
374
375
  input: readonly ["text", "file"];
375
376
  output: readonly ["text"];
376
377
  };
377
378
  capabilities: readonly ["attachments", "tool_call", "temperature"];
378
379
  context: number;
379
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
380
- }>, import("../../utils/preset").Preset<"meta/llama-3.1-70b", CatalogModel, {
380
+ }>, import("../../utils").Preset<"meta/llama-3.1-70b", CatalogModel, {
381
381
  name: string;
382
382
  created: string;
383
383
  knowledge: string;
384
+ providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
384
385
  modalities: {
385
386
  input: readonly ["text", "file"];
386
387
  output: readonly ["text"];
387
388
  };
388
389
  capabilities: readonly ["attachments", "tool_call", "temperature"];
389
390
  context: number;
390
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
391
- }>, import("../../utils/preset").Preset<"meta/llama-3.1-405b", CatalogModel, {
391
+ }>, import("../../utils").Preset<"meta/llama-3.1-405b", CatalogModel, {
392
392
  name: string;
393
393
  created: string;
394
394
  knowledge: string;
395
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "togetherai" | "vertex")[];
395
396
  modalities: {
396
397
  input: readonly ["text", "file"];
397
398
  output: readonly ["text"];
398
399
  };
399
400
  capabilities: readonly ["attachments", "tool_call", "temperature"];
400
401
  context: number;
401
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
402
402
  }>];
403
- readonly "v3.2": readonly [import("../../utils/preset").Preset<"meta/llama-3.2-1b", CatalogModel, {
403
+ readonly "v3.2": readonly [import("../../utils").Preset<"meta/llama-3.2-1b", CatalogModel, {
404
404
  name: string;
405
405
  created: string;
406
406
  knowledge: string;
@@ -410,74 +410,74 @@ export declare const llama: {
410
410
  };
411
411
  capabilities: readonly ["attachments", "tool_call", "temperature"];
412
412
  context: number;
413
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
414
- }>, import("../../utils/preset").Preset<"meta/llama-3.2-3b", CatalogModel, {
413
+ providers: readonly ["bedrock", "vertex", "azure"];
414
+ }>, import("../../utils").Preset<"meta/llama-3.2-3b", CatalogModel, {
415
415
  name: string;
416
416
  created: string;
417
417
  knowledge: string;
418
+ providers: ("azure" | "bedrock" | "fireworks" | "togetherai" | "vertex")[];
418
419
  modalities: {
419
420
  input: readonly ["text", "file"];
420
421
  output: readonly ["text"];
421
422
  };
422
423
  capabilities: readonly ["attachments", "tool_call", "temperature"];
423
424
  context: number;
424
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
425
- }>, import("../../utils/preset").Preset<"meta/llama-3.2-11b", CatalogModel, {
425
+ }>, import("../../utils").Preset<"meta/llama-3.2-11b", CatalogModel, {
426
426
  name: string;
427
427
  created: string;
428
428
  knowledge: string;
429
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "vertex")[];
429
430
  modalities: {
430
431
  input: readonly ["text", "file"];
431
432
  output: readonly ["text"];
432
433
  };
433
434
  capabilities: readonly ["attachments", "tool_call", "temperature"];
434
435
  context: number;
435
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
436
- }>, import("../../utils/preset").Preset<"meta/llama-3.2-90b", CatalogModel, {
436
+ }>, import("../../utils").Preset<"meta/llama-3.2-90b", CatalogModel, {
437
437
  name: string;
438
438
  created: string;
439
439
  knowledge: string;
440
+ providers: ("azure" | "bedrock" | "deepinfra" | "vertex")[];
440
441
  modalities: {
441
442
  input: readonly ["text", "file"];
442
443
  output: readonly ["text"];
443
444
  };
444
445
  capabilities: readonly ["attachments", "tool_call", "temperature"];
445
446
  context: number;
446
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
447
447
  }>];
448
- readonly "v3.3": readonly [import("../../utils/preset").Preset<"meta/llama-3.3-70b", CatalogModel, {
448
+ readonly "v3.3": readonly [import("../../utils").Preset<"meta/llama-3.3-70b", CatalogModel, {
449
449
  name: string;
450
450
  created: string;
451
451
  knowledge: string;
452
+ providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
452
453
  modalities: {
453
454
  input: readonly ["text", "file"];
454
455
  output: readonly ["text"];
455
456
  };
456
457
  capabilities: readonly ["attachments", "tool_call", "temperature"];
457
458
  context: number;
458
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
459
459
  }>];
460
- readonly v4: readonly [import("../../utils/preset").Preset<"meta/llama-4-scout", CatalogModel, {
460
+ readonly v4: readonly [import("../../utils").Preset<"meta/llama-4-scout", CatalogModel, {
461
461
  name: string;
462
462
  created: string;
463
463
  knowledge: string;
464
+ providers: ("azure" | "bedrock" | "deepinfra" | "groq" | "vertex")[];
464
465
  modalities: {
465
466
  input: readonly ["text", "image", "file"];
466
467
  output: readonly ["text"];
467
468
  };
468
469
  capabilities: readonly ["attachments", "tool_call", "temperature"];
469
470
  context: number;
470
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
471
- }>, import("../../utils/preset").Preset<"meta/llama-4-maverick", CatalogModel, {
471
+ }>, import("../../utils").Preset<"meta/llama-4-maverick", CatalogModel, {
472
472
  name: string;
473
473
  created: string;
474
474
  knowledge: string;
475
+ providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
475
476
  modalities: {
476
477
  input: readonly ["text", "image", "file"];
477
478
  output: readonly ["text"];
478
479
  };
479
480
  capabilities: readonly ["attachments", "tool_call", "temperature"];
480
481
  context: number;
481
- providers: readonly ["groq", "bedrock", "vertex", "azure"];
482
482
  }>];
483
483
  };