@hebo-ai/gateway 0.10.7 → 0.11.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +97 -84
- package/dist/config.js +21 -4
- package/dist/endpoints/chat-completions/converters.js +6 -2
- package/dist/endpoints/chat-completions/handler.js +9 -5
- package/dist/endpoints/chat-completions/schema.d.ts +48 -4
- package/dist/endpoints/chat-completions/schema.js +1 -1
- package/dist/endpoints/conversations/handler.js +3 -3
- package/dist/endpoints/embeddings/handler.js +2 -2
- package/dist/endpoints/messages/converters.js +17 -21
- package/dist/endpoints/messages/handler.js +5 -5
- package/dist/endpoints/responses/handler.js +9 -5
- package/dist/index.d.ts +1 -0
- package/dist/index.js +1 -0
- package/dist/models/alibaba/index.d.ts +2 -0
- package/dist/models/alibaba/index.js +2 -0
- package/dist/models/alibaba/middleware.d.ts +2 -0
- package/dist/models/alibaba/middleware.js +31 -0
- package/dist/models/alibaba/presets.d.ts +956 -0
- package/dist/models/alibaba/presets.js +264 -0
- package/dist/models/amazon/presets.d.ts +31 -31
- package/dist/models/anthropic/middleware.js +0 -1
- package/dist/models/anthropic/presets.d.ts +68 -68
- package/dist/models/cohere/presets.d.ts +38 -38
- package/dist/models/deepseek/index.d.ts +2 -0
- package/dist/models/deepseek/index.js +2 -0
- package/dist/models/deepseek/middleware.d.ts +2 -0
- package/dist/models/deepseek/middleware.js +25 -0
- package/dist/models/deepseek/presets.d.ts +51 -0
- package/dist/models/deepseek/presets.js +33 -0
- package/dist/models/google/presets.d.ts +106 -106
- package/dist/models/google/presets.js +11 -5
- package/dist/models/meta/presets.d.ts +84 -84
- package/dist/models/meta/presets.js +11 -12
- package/dist/models/minimax/index.d.ts +2 -0
- package/dist/models/minimax/index.js +2 -0
- package/dist/models/minimax/middleware.d.ts +2 -0
- package/dist/models/minimax/middleware.js +43 -0
- package/dist/models/minimax/presets.d.ts +99 -0
- package/dist/models/minimax/presets.js +47 -0
- package/dist/models/moonshot/index.d.ts +2 -0
- package/dist/models/moonshot/index.js +2 -0
- package/dist/models/moonshot/middleware.d.ts +2 -0
- package/dist/models/moonshot/middleware.js +32 -0
- package/dist/models/moonshot/presets.d.ts +104 -0
- package/dist/models/moonshot/presets.js +36 -0
- package/dist/models/openai/presets.d.ts +128 -128
- package/dist/models/openai/presets.js +1 -1
- package/dist/models/types.d.ts +1 -1
- package/dist/models/types.js +41 -1
- package/dist/models/voyage/presets.d.ts +46 -46
- package/dist/models/xai/index.d.ts +2 -0
- package/dist/models/xai/index.js +2 -0
- package/dist/models/xai/middleware.d.ts +2 -0
- package/dist/models/xai/middleware.js +37 -0
- package/dist/models/xai/presets.d.ts +210 -0
- package/dist/models/xai/presets.js +55 -0
- package/dist/models/zai/index.d.ts +2 -0
- package/dist/models/zai/index.js +2 -0
- package/dist/models/zai/middleware.d.ts +2 -0
- package/dist/models/zai/middleware.js +25 -0
- package/dist/models/zai/presets.d.ts +141 -0
- package/dist/models/zai/presets.js +41 -0
- package/dist/providers/alibaba/canonical.d.ts +3 -0
- package/dist/providers/alibaba/canonical.js +13 -0
- package/dist/providers/alibaba/index.d.ts +1 -0
- package/dist/providers/alibaba/index.js +1 -0
- package/dist/providers/bedrock/canonical.js +3 -0
- package/dist/providers/chutes/canonical.d.ts +3 -0
- package/dist/providers/chutes/canonical.js +14 -0
- package/dist/providers/chutes/index.d.ts +1 -0
- package/dist/providers/chutes/index.js +1 -0
- package/dist/providers/deepinfra/canonical.d.ts +3 -0
- package/dist/providers/deepinfra/canonical.js +40 -0
- package/dist/providers/deepinfra/index.d.ts +1 -0
- package/dist/providers/deepinfra/index.js +1 -0
- package/dist/providers/deepseek/canonical.d.ts +3 -0
- package/dist/providers/deepseek/canonical.js +9 -0
- package/dist/providers/deepseek/index.d.ts +1 -0
- package/dist/providers/deepseek/index.js +1 -0
- package/dist/providers/fireworks/canonical.d.ts +3 -0
- package/dist/providers/fireworks/canonical.js +27 -0
- package/dist/providers/fireworks/index.d.ts +2 -0
- package/dist/providers/fireworks/index.js +2 -0
- package/dist/providers/fireworks/middleware.d.ts +2 -0
- package/dist/providers/fireworks/middleware.js +35 -0
- package/dist/providers/groq/canonical.js +1 -1
- package/dist/providers/minimax/canonical.d.ts +3 -0
- package/dist/providers/minimax/canonical.js +9 -0
- package/dist/providers/minimax/index.d.ts +1 -0
- package/dist/providers/minimax/index.js +1 -0
- package/dist/providers/moonshot/canonical.d.ts +3 -0
- package/dist/providers/moonshot/canonical.js +6 -0
- package/dist/providers/moonshot/index.d.ts +1 -0
- package/dist/providers/moonshot/index.js +1 -0
- package/dist/providers/togetherai/canonical.d.ts +3 -0
- package/dist/providers/togetherai/canonical.js +24 -0
- package/dist/providers/togetherai/index.d.ts +1 -0
- package/dist/providers/togetherai/index.js +1 -0
- package/dist/providers/types.d.ts +1 -1
- package/dist/providers/types.js +10 -0
- package/dist/providers/vertex/canonical.js +5 -1
- package/dist/providers/xai/canonical.d.ts +3 -0
- package/dist/providers/xai/canonical.js +12 -0
- package/dist/providers/xai/index.d.ts +1 -0
- package/dist/providers/xai/index.js +1 -0
- package/dist/providers/zai/canonical.d.ts +3 -0
- package/dist/providers/zai/canonical.js +10 -0
- package/dist/providers/zai/index.d.ts +1 -0
- package/dist/providers/zai/index.js +1 -0
- package/dist/types.d.ts +30 -16
- package/dist/utils/index.d.ts +8 -0
- package/dist/utils/index.js +8 -0
- package/dist/utils/request.d.ts +2 -1
- package/dist/utils/request.js +35 -7
- package/package.json +154 -30
|
@@ -27,10 +27,14 @@ export const geminiEmbedding001 = presetFor()("google/embedding-001", {
|
|
|
27
27
|
created: "2025-05-20",
|
|
28
28
|
context: 8192,
|
|
29
29
|
});
|
|
30
|
-
export const
|
|
30
|
+
export const geminiEmbedding2 = presetFor()("google/gemini-embedding-2", {
|
|
31
31
|
...GEMINI_EMBEDDINGS_BASE,
|
|
32
|
-
|
|
33
|
-
|
|
32
|
+
modalities: {
|
|
33
|
+
input: ["text", "image", "video", "audio", "pdf"],
|
|
34
|
+
output: ["embedding"],
|
|
35
|
+
},
|
|
36
|
+
name: "Gemini Embedding 2",
|
|
37
|
+
created: "2026-04-23",
|
|
34
38
|
context: 8192,
|
|
35
39
|
});
|
|
36
40
|
export const gemini3FlashPreview = presetFor()("google/gemini-3-flash-preview", {
|
|
@@ -80,7 +84,7 @@ const GEMMA3_BASE = {
|
|
|
80
84
|
capabilities: ["tool_call", "structured_output", "temperature"],
|
|
81
85
|
context: 131072,
|
|
82
86
|
knowledge: "2025-01",
|
|
83
|
-
providers: ["vertex", "bedrock"],
|
|
87
|
+
providers: ["vertex", "bedrock", "deepinfra"],
|
|
84
88
|
};
|
|
85
89
|
const GEMMA4_BASE = {
|
|
86
90
|
modalities: {
|
|
@@ -141,12 +145,14 @@ export const gemma426bA4b = presetFor()("google/gemma-4-26b-a4b", {
|
|
|
141
145
|
name: "Gemma 4 26B-A4B",
|
|
142
146
|
created: "2026-04-02",
|
|
143
147
|
context: 262144,
|
|
148
|
+
providers: ["vertex", "deepinfra"],
|
|
144
149
|
});
|
|
145
150
|
export const gemma431b = presetFor()("google/gemma-4-31b", {
|
|
146
151
|
...GEMMA4_BASE,
|
|
147
152
|
name: "Gemma 4 31B",
|
|
148
153
|
created: "2026-04-02",
|
|
149
154
|
context: 262144,
|
|
155
|
+
providers: ["vertex", "deepinfra", "togetherai"],
|
|
150
156
|
});
|
|
151
157
|
const gemmaAtomic = {
|
|
152
158
|
v3: [gemma31b, gemma34b, gemma312b, gemma327b],
|
|
@@ -168,7 +174,7 @@ export const gemma = {
|
|
|
168
174
|
const geminiAtomic = {
|
|
169
175
|
"v2.5": [gemini25FlashLite, gemini25Flash, gemini25Pro],
|
|
170
176
|
"v3-preview": [gemini3FlashPreview, gemini31FlashLitePreview, gemini31ProPreview],
|
|
171
|
-
embeddings: [geminiEmbedding001,
|
|
177
|
+
embeddings: [geminiEmbedding001, geminiEmbedding2],
|
|
172
178
|
};
|
|
173
179
|
const geminiGroups = {
|
|
174
180
|
"v2.x": [...geminiAtomic["v2.5"]],
|
|
@@ -1,41 +1,41 @@
|
|
|
1
1
|
import type { CatalogModel } from "../types";
|
|
2
|
-
export declare const llama31_8b: import("../../utils
|
|
2
|
+
export declare const llama31_8b: import("../../utils").Preset<"meta/llama-3.1-8b", CatalogModel, {
|
|
3
3
|
name: string;
|
|
4
4
|
created: string;
|
|
5
5
|
knowledge: string;
|
|
6
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
|
|
6
7
|
modalities: {
|
|
7
8
|
input: readonly ["text", "file"];
|
|
8
9
|
output: readonly ["text"];
|
|
9
10
|
};
|
|
10
11
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
11
12
|
context: number;
|
|
12
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
13
13
|
}>;
|
|
14
|
-
export declare const llama31_70b: import("../../utils
|
|
14
|
+
export declare const llama31_70b: import("../../utils").Preset<"meta/llama-3.1-70b", CatalogModel, {
|
|
15
15
|
name: string;
|
|
16
16
|
created: string;
|
|
17
17
|
knowledge: string;
|
|
18
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
|
|
18
19
|
modalities: {
|
|
19
20
|
input: readonly ["text", "file"];
|
|
20
21
|
output: readonly ["text"];
|
|
21
22
|
};
|
|
22
23
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
23
24
|
context: number;
|
|
24
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
25
25
|
}>;
|
|
26
|
-
export declare const llama31_405b: import("../../utils
|
|
26
|
+
export declare const llama31_405b: import("../../utils").Preset<"meta/llama-3.1-405b", CatalogModel, {
|
|
27
27
|
name: string;
|
|
28
28
|
created: string;
|
|
29
29
|
knowledge: string;
|
|
30
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "togetherai" | "vertex")[];
|
|
30
31
|
modalities: {
|
|
31
32
|
input: readonly ["text", "file"];
|
|
32
33
|
output: readonly ["text"];
|
|
33
34
|
};
|
|
34
35
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
35
36
|
context: number;
|
|
36
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
37
37
|
}>;
|
|
38
|
-
export declare const llama32_1b: import("../../utils
|
|
38
|
+
export declare const llama32_1b: import("../../utils").Preset<"meta/llama-3.2-1b", CatalogModel, {
|
|
39
39
|
name: string;
|
|
40
40
|
created: string;
|
|
41
41
|
knowledge: string;
|
|
@@ -45,138 +45,138 @@ export declare const llama32_1b: import("../../utils/preset").Preset<"meta/llama
|
|
|
45
45
|
};
|
|
46
46
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
47
47
|
context: number;
|
|
48
|
-
providers: readonly ["
|
|
48
|
+
providers: readonly ["bedrock", "vertex", "azure"];
|
|
49
49
|
}>;
|
|
50
|
-
export declare const llama32_3b: import("../../utils
|
|
50
|
+
export declare const llama32_3b: import("../../utils").Preset<"meta/llama-3.2-3b", CatalogModel, {
|
|
51
51
|
name: string;
|
|
52
52
|
created: string;
|
|
53
53
|
knowledge: string;
|
|
54
|
+
providers: ("azure" | "bedrock" | "fireworks" | "togetherai" | "vertex")[];
|
|
54
55
|
modalities: {
|
|
55
56
|
input: readonly ["text", "file"];
|
|
56
57
|
output: readonly ["text"];
|
|
57
58
|
};
|
|
58
59
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
59
60
|
context: number;
|
|
60
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
61
61
|
}>;
|
|
62
|
-
export declare const llama32_11b: import("../../utils
|
|
62
|
+
export declare const llama32_11b: import("../../utils").Preset<"meta/llama-3.2-11b", CatalogModel, {
|
|
63
63
|
name: string;
|
|
64
64
|
created: string;
|
|
65
65
|
knowledge: string;
|
|
66
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "vertex")[];
|
|
66
67
|
modalities: {
|
|
67
68
|
input: readonly ["text", "file"];
|
|
68
69
|
output: readonly ["text"];
|
|
69
70
|
};
|
|
70
71
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
71
72
|
context: number;
|
|
72
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
73
73
|
}>;
|
|
74
|
-
export declare const llama32_90b: import("../../utils
|
|
74
|
+
export declare const llama32_90b: import("../../utils").Preset<"meta/llama-3.2-90b", CatalogModel, {
|
|
75
75
|
name: string;
|
|
76
76
|
created: string;
|
|
77
77
|
knowledge: string;
|
|
78
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "vertex")[];
|
|
78
79
|
modalities: {
|
|
79
80
|
input: readonly ["text", "file"];
|
|
80
81
|
output: readonly ["text"];
|
|
81
82
|
};
|
|
82
83
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
83
84
|
context: number;
|
|
84
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
85
85
|
}>;
|
|
86
|
-
export declare const llama33_70b: import("../../utils
|
|
86
|
+
export declare const llama33_70b: import("../../utils").Preset<"meta/llama-3.3-70b", CatalogModel, {
|
|
87
87
|
name: string;
|
|
88
88
|
created: string;
|
|
89
89
|
knowledge: string;
|
|
90
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
|
|
90
91
|
modalities: {
|
|
91
92
|
input: readonly ["text", "file"];
|
|
92
93
|
output: readonly ["text"];
|
|
93
94
|
};
|
|
94
95
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
95
96
|
context: number;
|
|
96
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
97
97
|
}>;
|
|
98
|
-
export declare const llama4Scout: import("../../utils
|
|
98
|
+
export declare const llama4Scout: import("../../utils").Preset<"meta/llama-4-scout", CatalogModel, {
|
|
99
99
|
name: string;
|
|
100
100
|
created: string;
|
|
101
101
|
knowledge: string;
|
|
102
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "groq" | "vertex")[];
|
|
102
103
|
modalities: {
|
|
103
104
|
input: readonly ["text", "image", "file"];
|
|
104
105
|
output: readonly ["text"];
|
|
105
106
|
};
|
|
106
107
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
107
108
|
context: number;
|
|
108
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
109
109
|
}>;
|
|
110
|
-
export declare const llama4Maverick: import("../../utils
|
|
110
|
+
export declare const llama4Maverick: import("../../utils").Preset<"meta/llama-4-maverick", CatalogModel, {
|
|
111
111
|
name: string;
|
|
112
112
|
created: string;
|
|
113
113
|
knowledge: string;
|
|
114
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
|
|
114
115
|
modalities: {
|
|
115
116
|
input: readonly ["text", "image", "file"];
|
|
116
117
|
output: readonly ["text"];
|
|
117
118
|
};
|
|
118
119
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
119
120
|
context: number;
|
|
120
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
121
121
|
}>;
|
|
122
122
|
export declare const llama: {
|
|
123
|
-
readonly latest: readonly [import("../../utils
|
|
123
|
+
readonly latest: readonly [import("../../utils").Preset<"meta/llama-4-scout", CatalogModel, {
|
|
124
124
|
name: string;
|
|
125
125
|
created: string;
|
|
126
126
|
knowledge: string;
|
|
127
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "groq" | "vertex")[];
|
|
127
128
|
modalities: {
|
|
128
129
|
input: readonly ["text", "image", "file"];
|
|
129
130
|
output: readonly ["text"];
|
|
130
131
|
};
|
|
131
132
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
132
133
|
context: number;
|
|
133
|
-
|
|
134
|
-
}>, import("../../utils/preset").Preset<"meta/llama-4-maverick", CatalogModel, {
|
|
134
|
+
}>, import("../../utils").Preset<"meta/llama-4-maverick", CatalogModel, {
|
|
135
135
|
name: string;
|
|
136
136
|
created: string;
|
|
137
137
|
knowledge: string;
|
|
138
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
|
|
138
139
|
modalities: {
|
|
139
140
|
input: readonly ["text", "image", "file"];
|
|
140
141
|
output: readonly ["text"];
|
|
141
142
|
};
|
|
142
143
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
143
144
|
context: number;
|
|
144
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
145
145
|
}>];
|
|
146
|
-
readonly all: (import("../../utils
|
|
146
|
+
readonly all: (import("../../utils").Preset<"meta/llama-3.1-8b", CatalogModel, {
|
|
147
147
|
name: string;
|
|
148
148
|
created: string;
|
|
149
149
|
knowledge: string;
|
|
150
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
|
|
150
151
|
modalities: {
|
|
151
152
|
input: readonly ["text", "file"];
|
|
152
153
|
output: readonly ["text"];
|
|
153
154
|
};
|
|
154
155
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
155
156
|
context: number;
|
|
156
|
-
|
|
157
|
-
}> | import("../../utils/preset").Preset<"meta/llama-3.1-70b", CatalogModel, {
|
|
157
|
+
}> | import("../../utils").Preset<"meta/llama-3.1-70b", CatalogModel, {
|
|
158
158
|
name: string;
|
|
159
159
|
created: string;
|
|
160
160
|
knowledge: string;
|
|
161
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
|
|
161
162
|
modalities: {
|
|
162
163
|
input: readonly ["text", "file"];
|
|
163
164
|
output: readonly ["text"];
|
|
164
165
|
};
|
|
165
166
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
166
167
|
context: number;
|
|
167
|
-
|
|
168
|
-
}> | import("../../utils/preset").Preset<"meta/llama-3.1-405b", CatalogModel, {
|
|
168
|
+
}> | import("../../utils").Preset<"meta/llama-3.1-405b", CatalogModel, {
|
|
169
169
|
name: string;
|
|
170
170
|
created: string;
|
|
171
171
|
knowledge: string;
|
|
172
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "togetherai" | "vertex")[];
|
|
172
173
|
modalities: {
|
|
173
174
|
input: readonly ["text", "file"];
|
|
174
175
|
output: readonly ["text"];
|
|
175
176
|
};
|
|
176
177
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
177
178
|
context: number;
|
|
178
|
-
|
|
179
|
-
}> | import("../../utils/preset").Preset<"meta/llama-3.2-1b", CatalogModel, {
|
|
179
|
+
}> | import("../../utils").Preset<"meta/llama-3.2-1b", CatalogModel, {
|
|
180
180
|
name: string;
|
|
181
181
|
created: string;
|
|
182
182
|
knowledge: string;
|
|
@@ -186,108 +186,108 @@ export declare const llama: {
|
|
|
186
186
|
};
|
|
187
187
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
188
188
|
context: number;
|
|
189
|
-
providers: readonly ["
|
|
190
|
-
}> | import("../../utils
|
|
189
|
+
providers: readonly ["bedrock", "vertex", "azure"];
|
|
190
|
+
}> | import("../../utils").Preset<"meta/llama-3.2-3b", CatalogModel, {
|
|
191
191
|
name: string;
|
|
192
192
|
created: string;
|
|
193
193
|
knowledge: string;
|
|
194
|
+
providers: ("azure" | "bedrock" | "fireworks" | "togetherai" | "vertex")[];
|
|
194
195
|
modalities: {
|
|
195
196
|
input: readonly ["text", "file"];
|
|
196
197
|
output: readonly ["text"];
|
|
197
198
|
};
|
|
198
199
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
199
200
|
context: number;
|
|
200
|
-
|
|
201
|
-
}> | import("../../utils/preset").Preset<"meta/llama-3.2-11b", CatalogModel, {
|
|
201
|
+
}> | import("../../utils").Preset<"meta/llama-3.2-11b", CatalogModel, {
|
|
202
202
|
name: string;
|
|
203
203
|
created: string;
|
|
204
204
|
knowledge: string;
|
|
205
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "vertex")[];
|
|
205
206
|
modalities: {
|
|
206
207
|
input: readonly ["text", "file"];
|
|
207
208
|
output: readonly ["text"];
|
|
208
209
|
};
|
|
209
210
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
210
211
|
context: number;
|
|
211
|
-
|
|
212
|
-
}> | import("../../utils/preset").Preset<"meta/llama-3.2-90b", CatalogModel, {
|
|
212
|
+
}> | import("../../utils").Preset<"meta/llama-3.2-90b", CatalogModel, {
|
|
213
213
|
name: string;
|
|
214
214
|
created: string;
|
|
215
215
|
knowledge: string;
|
|
216
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "vertex")[];
|
|
216
217
|
modalities: {
|
|
217
218
|
input: readonly ["text", "file"];
|
|
218
219
|
output: readonly ["text"];
|
|
219
220
|
};
|
|
220
221
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
221
222
|
context: number;
|
|
222
|
-
|
|
223
|
-
}> | import("../../utils/preset").Preset<"meta/llama-3.3-70b", CatalogModel, {
|
|
223
|
+
}> | import("../../utils").Preset<"meta/llama-3.3-70b", CatalogModel, {
|
|
224
224
|
name: string;
|
|
225
225
|
created: string;
|
|
226
226
|
knowledge: string;
|
|
227
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
|
|
227
228
|
modalities: {
|
|
228
229
|
input: readonly ["text", "file"];
|
|
229
230
|
output: readonly ["text"];
|
|
230
231
|
};
|
|
231
232
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
232
233
|
context: number;
|
|
233
|
-
|
|
234
|
-
}> | import("../../utils/preset").Preset<"meta/llama-4-scout", CatalogModel, {
|
|
234
|
+
}> | import("../../utils").Preset<"meta/llama-4-scout", CatalogModel, {
|
|
235
235
|
name: string;
|
|
236
236
|
created: string;
|
|
237
237
|
knowledge: string;
|
|
238
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "groq" | "vertex")[];
|
|
238
239
|
modalities: {
|
|
239
240
|
input: readonly ["text", "image", "file"];
|
|
240
241
|
output: readonly ["text"];
|
|
241
242
|
};
|
|
242
243
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
243
244
|
context: number;
|
|
244
|
-
|
|
245
|
-
}> | import("../../utils/preset").Preset<"meta/llama-4-maverick", CatalogModel, {
|
|
245
|
+
}> | import("../../utils").Preset<"meta/llama-4-maverick", CatalogModel, {
|
|
246
246
|
name: string;
|
|
247
247
|
created: string;
|
|
248
248
|
knowledge: string;
|
|
249
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
|
|
249
250
|
modalities: {
|
|
250
251
|
input: readonly ["text", "image", "file"];
|
|
251
252
|
output: readonly ["text"];
|
|
252
253
|
};
|
|
253
254
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
254
255
|
context: number;
|
|
255
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
256
256
|
}>)[];
|
|
257
|
-
readonly "v3.x": readonly [import("../../utils
|
|
257
|
+
readonly "v3.x": readonly [import("../../utils").Preset<"meta/llama-3.1-8b", CatalogModel, {
|
|
258
258
|
name: string;
|
|
259
259
|
created: string;
|
|
260
260
|
knowledge: string;
|
|
261
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
|
|
261
262
|
modalities: {
|
|
262
263
|
input: readonly ["text", "file"];
|
|
263
264
|
output: readonly ["text"];
|
|
264
265
|
};
|
|
265
266
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
266
267
|
context: number;
|
|
267
|
-
|
|
268
|
-
}>, import("../../utils/preset").Preset<"meta/llama-3.1-70b", CatalogModel, {
|
|
268
|
+
}>, import("../../utils").Preset<"meta/llama-3.1-70b", CatalogModel, {
|
|
269
269
|
name: string;
|
|
270
270
|
created: string;
|
|
271
271
|
knowledge: string;
|
|
272
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
|
|
272
273
|
modalities: {
|
|
273
274
|
input: readonly ["text", "file"];
|
|
274
275
|
output: readonly ["text"];
|
|
275
276
|
};
|
|
276
277
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
277
278
|
context: number;
|
|
278
|
-
|
|
279
|
-
}>, import("../../utils/preset").Preset<"meta/llama-3.1-405b", CatalogModel, {
|
|
279
|
+
}>, import("../../utils").Preset<"meta/llama-3.1-405b", CatalogModel, {
|
|
280
280
|
name: string;
|
|
281
281
|
created: string;
|
|
282
282
|
knowledge: string;
|
|
283
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "togetherai" | "vertex")[];
|
|
283
284
|
modalities: {
|
|
284
285
|
input: readonly ["text", "file"];
|
|
285
286
|
output: readonly ["text"];
|
|
286
287
|
};
|
|
287
288
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
288
289
|
context: number;
|
|
289
|
-
|
|
290
|
-
}>, import("../../utils/preset").Preset<"meta/llama-3.2-1b", CatalogModel, {
|
|
290
|
+
}>, import("../../utils").Preset<"meta/llama-3.2-1b", CatalogModel, {
|
|
291
291
|
name: string;
|
|
292
292
|
created: string;
|
|
293
293
|
knowledge: string;
|
|
@@ -297,110 +297,110 @@ export declare const llama: {
|
|
|
297
297
|
};
|
|
298
298
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
299
299
|
context: number;
|
|
300
|
-
providers: readonly ["
|
|
301
|
-
}>, import("../../utils
|
|
300
|
+
providers: readonly ["bedrock", "vertex", "azure"];
|
|
301
|
+
}>, import("../../utils").Preset<"meta/llama-3.2-3b", CatalogModel, {
|
|
302
302
|
name: string;
|
|
303
303
|
created: string;
|
|
304
304
|
knowledge: string;
|
|
305
|
+
providers: ("azure" | "bedrock" | "fireworks" | "togetherai" | "vertex")[];
|
|
305
306
|
modalities: {
|
|
306
307
|
input: readonly ["text", "file"];
|
|
307
308
|
output: readonly ["text"];
|
|
308
309
|
};
|
|
309
310
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
310
311
|
context: number;
|
|
311
|
-
|
|
312
|
-
}>, import("../../utils/preset").Preset<"meta/llama-3.2-11b", CatalogModel, {
|
|
312
|
+
}>, import("../../utils").Preset<"meta/llama-3.2-11b", CatalogModel, {
|
|
313
313
|
name: string;
|
|
314
314
|
created: string;
|
|
315
315
|
knowledge: string;
|
|
316
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "vertex")[];
|
|
316
317
|
modalities: {
|
|
317
318
|
input: readonly ["text", "file"];
|
|
318
319
|
output: readonly ["text"];
|
|
319
320
|
};
|
|
320
321
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
321
322
|
context: number;
|
|
322
|
-
|
|
323
|
-
}>, import("../../utils/preset").Preset<"meta/llama-3.2-90b", CatalogModel, {
|
|
323
|
+
}>, import("../../utils").Preset<"meta/llama-3.2-90b", CatalogModel, {
|
|
324
324
|
name: string;
|
|
325
325
|
created: string;
|
|
326
326
|
knowledge: string;
|
|
327
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "vertex")[];
|
|
327
328
|
modalities: {
|
|
328
329
|
input: readonly ["text", "file"];
|
|
329
330
|
output: readonly ["text"];
|
|
330
331
|
};
|
|
331
332
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
332
333
|
context: number;
|
|
333
|
-
|
|
334
|
-
}>, import("../../utils/preset").Preset<"meta/llama-3.3-70b", CatalogModel, {
|
|
334
|
+
}>, import("../../utils").Preset<"meta/llama-3.3-70b", CatalogModel, {
|
|
335
335
|
name: string;
|
|
336
336
|
created: string;
|
|
337
337
|
knowledge: string;
|
|
338
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
|
|
338
339
|
modalities: {
|
|
339
340
|
input: readonly ["text", "file"];
|
|
340
341
|
output: readonly ["text"];
|
|
341
342
|
};
|
|
342
343
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
343
344
|
context: number;
|
|
344
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
345
345
|
}>];
|
|
346
|
-
readonly "v4.x": readonly [import("../../utils
|
|
346
|
+
readonly "v4.x": readonly [import("../../utils").Preset<"meta/llama-4-scout", CatalogModel, {
|
|
347
347
|
name: string;
|
|
348
348
|
created: string;
|
|
349
349
|
knowledge: string;
|
|
350
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "groq" | "vertex")[];
|
|
350
351
|
modalities: {
|
|
351
352
|
input: readonly ["text", "image", "file"];
|
|
352
353
|
output: readonly ["text"];
|
|
353
354
|
};
|
|
354
355
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
355
356
|
context: number;
|
|
356
|
-
|
|
357
|
-
}>, import("../../utils/preset").Preset<"meta/llama-4-maverick", CatalogModel, {
|
|
357
|
+
}>, import("../../utils").Preset<"meta/llama-4-maverick", CatalogModel, {
|
|
358
358
|
name: string;
|
|
359
359
|
created: string;
|
|
360
360
|
knowledge: string;
|
|
361
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
|
|
361
362
|
modalities: {
|
|
362
363
|
input: readonly ["text", "image", "file"];
|
|
363
364
|
output: readonly ["text"];
|
|
364
365
|
};
|
|
365
366
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
366
367
|
context: number;
|
|
367
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
368
368
|
}>];
|
|
369
|
-
readonly "v3.1": readonly [import("../../utils
|
|
369
|
+
readonly "v3.1": readonly [import("../../utils").Preset<"meta/llama-3.1-8b", CatalogModel, {
|
|
370
370
|
name: string;
|
|
371
371
|
created: string;
|
|
372
372
|
knowledge: string;
|
|
373
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
|
|
373
374
|
modalities: {
|
|
374
375
|
input: readonly ["text", "file"];
|
|
375
376
|
output: readonly ["text"];
|
|
376
377
|
};
|
|
377
378
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
378
379
|
context: number;
|
|
379
|
-
|
|
380
|
-
}>, import("../../utils/preset").Preset<"meta/llama-3.1-70b", CatalogModel, {
|
|
380
|
+
}>, import("../../utils").Preset<"meta/llama-3.1-70b", CatalogModel, {
|
|
381
381
|
name: string;
|
|
382
382
|
created: string;
|
|
383
383
|
knowledge: string;
|
|
384
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
|
|
384
385
|
modalities: {
|
|
385
386
|
input: readonly ["text", "file"];
|
|
386
387
|
output: readonly ["text"];
|
|
387
388
|
};
|
|
388
389
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
389
390
|
context: number;
|
|
390
|
-
|
|
391
|
-
}>, import("../../utils/preset").Preset<"meta/llama-3.1-405b", CatalogModel, {
|
|
391
|
+
}>, import("../../utils").Preset<"meta/llama-3.1-405b", CatalogModel, {
|
|
392
392
|
name: string;
|
|
393
393
|
created: string;
|
|
394
394
|
knowledge: string;
|
|
395
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "togetherai" | "vertex")[];
|
|
395
396
|
modalities: {
|
|
396
397
|
input: readonly ["text", "file"];
|
|
397
398
|
output: readonly ["text"];
|
|
398
399
|
};
|
|
399
400
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
400
401
|
context: number;
|
|
401
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
402
402
|
}>];
|
|
403
|
-
readonly "v3.2": readonly [import("../../utils
|
|
403
|
+
readonly "v3.2": readonly [import("../../utils").Preset<"meta/llama-3.2-1b", CatalogModel, {
|
|
404
404
|
name: string;
|
|
405
405
|
created: string;
|
|
406
406
|
knowledge: string;
|
|
@@ -410,74 +410,74 @@ export declare const llama: {
|
|
|
410
410
|
};
|
|
411
411
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
412
412
|
context: number;
|
|
413
|
-
providers: readonly ["
|
|
414
|
-
}>, import("../../utils
|
|
413
|
+
providers: readonly ["bedrock", "vertex", "azure"];
|
|
414
|
+
}>, import("../../utils").Preset<"meta/llama-3.2-3b", CatalogModel, {
|
|
415
415
|
name: string;
|
|
416
416
|
created: string;
|
|
417
417
|
knowledge: string;
|
|
418
|
+
providers: ("azure" | "bedrock" | "fireworks" | "togetherai" | "vertex")[];
|
|
418
419
|
modalities: {
|
|
419
420
|
input: readonly ["text", "file"];
|
|
420
421
|
output: readonly ["text"];
|
|
421
422
|
};
|
|
422
423
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
423
424
|
context: number;
|
|
424
|
-
|
|
425
|
-
}>, import("../../utils/preset").Preset<"meta/llama-3.2-11b", CatalogModel, {
|
|
425
|
+
}>, import("../../utils").Preset<"meta/llama-3.2-11b", CatalogModel, {
|
|
426
426
|
name: string;
|
|
427
427
|
created: string;
|
|
428
428
|
knowledge: string;
|
|
429
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "vertex")[];
|
|
429
430
|
modalities: {
|
|
430
431
|
input: readonly ["text", "file"];
|
|
431
432
|
output: readonly ["text"];
|
|
432
433
|
};
|
|
433
434
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
434
435
|
context: number;
|
|
435
|
-
|
|
436
|
-
}>, import("../../utils/preset").Preset<"meta/llama-3.2-90b", CatalogModel, {
|
|
436
|
+
}>, import("../../utils").Preset<"meta/llama-3.2-90b", CatalogModel, {
|
|
437
437
|
name: string;
|
|
438
438
|
created: string;
|
|
439
439
|
knowledge: string;
|
|
440
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "vertex")[];
|
|
440
441
|
modalities: {
|
|
441
442
|
input: readonly ["text", "file"];
|
|
442
443
|
output: readonly ["text"];
|
|
443
444
|
};
|
|
444
445
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
445
446
|
context: number;
|
|
446
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
447
447
|
}>];
|
|
448
|
-
readonly "v3.3": readonly [import("../../utils
|
|
448
|
+
readonly "v3.3": readonly [import("../../utils").Preset<"meta/llama-3.3-70b", CatalogModel, {
|
|
449
449
|
name: string;
|
|
450
450
|
created: string;
|
|
451
451
|
knowledge: string;
|
|
452
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "fireworks" | "groq" | "togetherai" | "vertex")[];
|
|
452
453
|
modalities: {
|
|
453
454
|
input: readonly ["text", "file"];
|
|
454
455
|
output: readonly ["text"];
|
|
455
456
|
};
|
|
456
457
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
457
458
|
context: number;
|
|
458
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
459
459
|
}>];
|
|
460
|
-
readonly v4: readonly [import("../../utils
|
|
460
|
+
readonly v4: readonly [import("../../utils").Preset<"meta/llama-4-scout", CatalogModel, {
|
|
461
461
|
name: string;
|
|
462
462
|
created: string;
|
|
463
463
|
knowledge: string;
|
|
464
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "groq" | "vertex")[];
|
|
464
465
|
modalities: {
|
|
465
466
|
input: readonly ["text", "image", "file"];
|
|
466
467
|
output: readonly ["text"];
|
|
467
468
|
};
|
|
468
469
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
469
470
|
context: number;
|
|
470
|
-
|
|
471
|
-
}>, import("../../utils/preset").Preset<"meta/llama-4-maverick", CatalogModel, {
|
|
471
|
+
}>, import("../../utils").Preset<"meta/llama-4-maverick", CatalogModel, {
|
|
472
472
|
name: string;
|
|
473
473
|
created: string;
|
|
474
474
|
knowledge: string;
|
|
475
|
+
providers: ("azure" | "bedrock" | "deepinfra" | "togetherai" | "vertex")[];
|
|
475
476
|
modalities: {
|
|
476
477
|
input: readonly ["text", "image", "file"];
|
|
477
478
|
output: readonly ["text"];
|
|
478
479
|
};
|
|
479
480
|
capabilities: readonly ["attachments", "tool_call", "temperature"];
|
|
480
481
|
context: number;
|
|
481
|
-
providers: readonly ["groq", "bedrock", "vertex", "azure"];
|
|
482
482
|
}>];
|
|
483
483
|
};
|