@chainfuse/types 1.6.7 → 1.6.9
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
|
@@ -11,6 +11,7 @@ export declare const workersAiCatalog: {
|
|
|
11
11
|
readonly tags: readonly [];
|
|
12
12
|
readonly properties: {
|
|
13
13
|
readonly beta: true;
|
|
14
|
+
readonly context_window: 32000;
|
|
14
15
|
readonly info: "https://huggingface.co/qwen/qwen1.5-0.5b-chat";
|
|
15
16
|
};
|
|
16
17
|
}, {
|
|
@@ -21,6 +22,7 @@ export declare const workersAiCatalog: {
|
|
|
21
22
|
readonly tags: readonly [];
|
|
22
23
|
readonly properties: {
|
|
23
24
|
readonly beta: true;
|
|
25
|
+
readonly context_window: 8192;
|
|
24
26
|
readonly lora: true;
|
|
25
27
|
};
|
|
26
28
|
}, {
|
|
@@ -31,6 +33,7 @@ export declare const workersAiCatalog: {
|
|
|
31
33
|
readonly tags: readonly [];
|
|
32
34
|
readonly properties: {
|
|
33
35
|
readonly beta: true;
|
|
36
|
+
readonly context_window: 4096;
|
|
34
37
|
readonly info: "https://huggingface.co/Nexusflow/Starling-LM-7B-beta";
|
|
35
38
|
readonly max_batch_prefill_tokens: 8192;
|
|
36
39
|
readonly max_input_length: 3072;
|
|
@@ -43,6 +46,7 @@ export declare const workersAiCatalog: {
|
|
|
43
46
|
readonly description: "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.";
|
|
44
47
|
readonly tags: readonly [];
|
|
45
48
|
readonly properties: {
|
|
49
|
+
readonly context_window: 7968;
|
|
46
50
|
readonly info: "https://llama.meta.com";
|
|
47
51
|
readonly terms: "https://llama.meta.com/llama3/license/#";
|
|
48
52
|
};
|
|
@@ -53,6 +57,7 @@ export declare const workersAiCatalog: {
|
|
|
53
57
|
readonly description: "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
|
|
54
58
|
readonly tags: readonly [];
|
|
55
59
|
readonly properties: {
|
|
60
|
+
readonly context_window: 128000;
|
|
56
61
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
|
|
57
62
|
};
|
|
58
63
|
}, {
|
|
@@ -63,6 +68,7 @@ export declare const workersAiCatalog: {
|
|
|
63
68
|
readonly tags: readonly [];
|
|
64
69
|
readonly properties: {
|
|
65
70
|
readonly beta: true;
|
|
71
|
+
readonly context_window: 4096;
|
|
66
72
|
};
|
|
67
73
|
}, {
|
|
68
74
|
readonly id: "d2ba5c6b-bbb7-49d6-b466-900654870cd6";
|
|
@@ -72,7 +78,15 @@ export declare const workersAiCatalog: {
|
|
|
72
78
|
readonly tags: readonly [];
|
|
73
79
|
readonly properties: {
|
|
74
80
|
readonly beta: true;
|
|
81
|
+
readonly context_window: 4096;
|
|
75
82
|
};
|
|
83
|
+
}, {
|
|
84
|
+
readonly id: "cc80437b-9a8d-4f1a-9c77-9aaf0d226922";
|
|
85
|
+
readonly source: 1;
|
|
86
|
+
readonly name: "@cf/meta/llama-guard-3-8b";
|
|
87
|
+
readonly description: "Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.";
|
|
88
|
+
readonly tags: readonly [];
|
|
89
|
+
readonly properties: {};
|
|
76
90
|
}, {
|
|
77
91
|
readonly id: "ca54bcd6-0d98-4739-9b3b-5c8b4402193d";
|
|
78
92
|
readonly source: 1;
|
|
@@ -81,6 +95,7 @@ export declare const workersAiCatalog: {
|
|
|
81
95
|
readonly tags: readonly [];
|
|
82
96
|
readonly properties: {
|
|
83
97
|
readonly beta: true;
|
|
98
|
+
readonly context_window: 4096;
|
|
84
99
|
readonly info: "https://ai.meta.com/llama/";
|
|
85
100
|
readonly terms: "https://ai.meta.com/resources/models-and-libraries/llama-downloads/";
|
|
86
101
|
};
|
|
@@ -92,6 +107,7 @@ export declare const workersAiCatalog: {
|
|
|
92
107
|
readonly tags: readonly [];
|
|
93
108
|
readonly properties: {
|
|
94
109
|
readonly beta: true;
|
|
110
|
+
readonly context_window: 2824;
|
|
95
111
|
readonly info: "https://mistral.ai/news/announcing-mistral-7b/";
|
|
96
112
|
readonly lora: true;
|
|
97
113
|
};
|
|
@@ -103,6 +119,7 @@ export declare const workersAiCatalog: {
|
|
|
103
119
|
readonly tags: readonly [];
|
|
104
120
|
readonly properties: {
|
|
105
121
|
readonly beta: true;
|
|
122
|
+
readonly context_window: 15000;
|
|
106
123
|
readonly lora: true;
|
|
107
124
|
};
|
|
108
125
|
}, {
|
|
@@ -113,6 +130,7 @@ export declare const workersAiCatalog: {
|
|
|
113
130
|
readonly tags: readonly [];
|
|
114
131
|
readonly properties: {
|
|
115
132
|
readonly beta: true;
|
|
133
|
+
readonly context_window: 2048;
|
|
116
134
|
readonly info: "https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0";
|
|
117
135
|
};
|
|
118
136
|
}, {
|
|
@@ -123,6 +141,7 @@ export declare const workersAiCatalog: {
|
|
|
123
141
|
readonly tags: readonly [];
|
|
124
142
|
readonly properties: {
|
|
125
143
|
readonly beta: true;
|
|
144
|
+
readonly context_window: 3072;
|
|
126
145
|
readonly info: "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2";
|
|
127
146
|
readonly lora: true;
|
|
128
147
|
readonly max_batch_prefill_tokens: 8192;
|
|
@@ -137,6 +156,7 @@ export declare const workersAiCatalog: {
|
|
|
137
156
|
readonly tags: readonly [];
|
|
138
157
|
readonly properties: {
|
|
139
158
|
readonly beta: true;
|
|
159
|
+
readonly context_window: 15000;
|
|
140
160
|
};
|
|
141
161
|
}, {
|
|
142
162
|
readonly id: "ad01ab83-baf8-4e7b-8fed-a0a219d4eb45";
|
|
@@ -145,6 +165,7 @@ export declare const workersAiCatalog: {
|
|
|
145
165
|
readonly description: "DeepSeek-R1-Distill-Qwen-32B is a model distilled from DeepSeek-R1 based on Qwen2.5. It outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.";
|
|
146
166
|
readonly tags: readonly [];
|
|
147
167
|
readonly properties: {
|
|
168
|
+
readonly context_window: 80000;
|
|
148
169
|
readonly terms: "https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE";
|
|
149
170
|
};
|
|
150
171
|
}, {
|
|
@@ -155,6 +176,7 @@ export declare const workersAiCatalog: {
|
|
|
155
176
|
readonly tags: readonly [];
|
|
156
177
|
readonly properties: {
|
|
157
178
|
readonly beta: true;
|
|
179
|
+
readonly context_window: 4096;
|
|
158
180
|
readonly info: "https://huggingface.co/TheBloke/DiscoLM_German_7b_v1-AWQ";
|
|
159
181
|
};
|
|
160
182
|
}, {
|
|
@@ -163,7 +185,9 @@ export declare const workersAiCatalog: {
|
|
|
163
185
|
readonly name: "@cf/meta/llama-2-7b-chat-int8";
|
|
164
186
|
readonly description: "Quantized (int8) generative text model with 7 billion parameters from Meta";
|
|
165
187
|
readonly tags: readonly [];
|
|
166
|
-
readonly properties: {
|
|
188
|
+
readonly properties: {
|
|
189
|
+
readonly context_window: 8192;
|
|
190
|
+
};
|
|
167
191
|
}, {
|
|
168
192
|
readonly id: "9b9c87c6-d4b7-494c-b177-87feab5904db";
|
|
169
193
|
readonly source: 1;
|
|
@@ -171,6 +195,7 @@ export declare const workersAiCatalog: {
|
|
|
171
195
|
readonly description: "Llama 3.1 8B quantized to FP8 precision";
|
|
172
196
|
readonly tags: readonly [];
|
|
173
197
|
readonly properties: {
|
|
198
|
+
readonly context_window: 32000;
|
|
174
199
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
|
|
175
200
|
};
|
|
176
201
|
}, {
|
|
@@ -181,6 +206,7 @@ export declare const workersAiCatalog: {
|
|
|
181
206
|
readonly tags: readonly [];
|
|
182
207
|
readonly properties: {
|
|
183
208
|
readonly beta: true;
|
|
209
|
+
readonly context_window: 4096;
|
|
184
210
|
readonly info: "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-AWQ";
|
|
185
211
|
};
|
|
186
212
|
}, {
|
|
@@ -191,6 +217,7 @@ export declare const workersAiCatalog: {
|
|
|
191
217
|
readonly tags: readonly [];
|
|
192
218
|
readonly properties: {
|
|
193
219
|
readonly beta: true;
|
|
220
|
+
readonly context_window: 20000;
|
|
194
221
|
readonly info: "https://huggingface.co/qwen/qwen1.5-7b-chat-awq";
|
|
195
222
|
};
|
|
196
223
|
}, {
|
|
@@ -200,6 +227,7 @@ export declare const workersAiCatalog: {
|
|
|
200
227
|
readonly description: "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
|
|
201
228
|
readonly tags: readonly [];
|
|
202
229
|
readonly properties: {
|
|
230
|
+
readonly context_window: 128000;
|
|
203
231
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
|
|
204
232
|
};
|
|
205
233
|
}, {
|
|
@@ -210,6 +238,7 @@ export declare const workersAiCatalog: {
|
|
|
210
238
|
readonly tags: readonly [];
|
|
211
239
|
readonly properties: {
|
|
212
240
|
readonly beta: true;
|
|
241
|
+
readonly context_window: 4096;
|
|
213
242
|
readonly info: "https://huggingface.co/TheBloke/Llama-2-13B-chat-AWQ";
|
|
214
243
|
};
|
|
215
244
|
}, {
|
|
@@ -220,6 +249,7 @@ export declare const workersAiCatalog: {
|
|
|
220
249
|
readonly tags: readonly [];
|
|
221
250
|
readonly properties: {
|
|
222
251
|
readonly beta: true;
|
|
252
|
+
readonly context_window: 4096;
|
|
223
253
|
readonly terms: "https://huggingface.co/TheBloke/deepseek-coder-6.7B-base-AWQ";
|
|
224
254
|
};
|
|
225
255
|
}, {
|
|
@@ -230,6 +260,7 @@ export declare const workersAiCatalog: {
|
|
|
230
260
|
readonly tags: readonly [];
|
|
231
261
|
readonly properties: {
|
|
232
262
|
readonly beta: true;
|
|
263
|
+
readonly context_window: 8192;
|
|
233
264
|
readonly lora: true;
|
|
234
265
|
};
|
|
235
266
|
}, {
|
|
@@ -239,6 +270,7 @@ export declare const workersAiCatalog: {
|
|
|
239
270
|
readonly description: "Llama 3.3 70B quantized to fp8 precision, optimized to be faster.";
|
|
240
271
|
readonly tags: readonly [];
|
|
241
272
|
readonly properties: {
|
|
273
|
+
readonly context_window: 24000;
|
|
242
274
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_3/LICENSE";
|
|
243
275
|
};
|
|
244
276
|
}, {
|
|
@@ -249,6 +281,7 @@ export declare const workersAiCatalog: {
|
|
|
249
281
|
readonly tags: readonly [];
|
|
250
282
|
readonly properties: {
|
|
251
283
|
readonly beta: true;
|
|
284
|
+
readonly context_window: 4096;
|
|
252
285
|
};
|
|
253
286
|
}, {
|
|
254
287
|
readonly id: "60474554-f03b-4ff4-8ecc-c1b7c71d7b29";
|
|
@@ -258,6 +291,7 @@ export declare const workersAiCatalog: {
|
|
|
258
291
|
readonly tags: readonly [];
|
|
259
292
|
readonly properties: {
|
|
260
293
|
readonly beta: true;
|
|
294
|
+
readonly context_window: 4096;
|
|
261
295
|
readonly terms: "https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-AWQ";
|
|
262
296
|
};
|
|
263
297
|
}, {
|
|
@@ -268,6 +302,7 @@ export declare const workersAiCatalog: {
|
|
|
268
302
|
readonly tags: readonly [];
|
|
269
303
|
readonly properties: {
|
|
270
304
|
readonly beta: true;
|
|
305
|
+
readonly context_window: 4096;
|
|
271
306
|
readonly info: "https://huggingface.co/deepseek-ai/deepseek-math-7b-instruct";
|
|
272
307
|
readonly terms: "https://github.com/deepseek-ai/DeepSeek-Math/blob/main/LICENSE-MODEL";
|
|
273
308
|
};
|
|
@@ -279,6 +314,7 @@ export declare const workersAiCatalog: {
|
|
|
279
314
|
readonly tags: readonly [];
|
|
280
315
|
readonly properties: {
|
|
281
316
|
readonly beta: true;
|
|
317
|
+
readonly context_window: 4096;
|
|
282
318
|
readonly info: "https://huggingface.co/tiiuae/falcon-7b-instruct";
|
|
283
319
|
};
|
|
284
320
|
}, {
|
|
@@ -289,6 +325,7 @@ export declare const workersAiCatalog: {
|
|
|
289
325
|
readonly tags: readonly [];
|
|
290
326
|
readonly properties: {
|
|
291
327
|
readonly beta: true;
|
|
328
|
+
readonly context_window: 24000;
|
|
292
329
|
readonly function_calling: true;
|
|
293
330
|
readonly info: "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B";
|
|
294
331
|
};
|
|
@@ -299,6 +336,7 @@ export declare const workersAiCatalog: {
|
|
|
299
336
|
readonly description: "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.";
|
|
300
337
|
readonly tags: readonly [];
|
|
301
338
|
readonly properties: {
|
|
339
|
+
readonly context_window: 7968;
|
|
302
340
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
|
|
303
341
|
};
|
|
304
342
|
}, {
|
|
@@ -308,6 +346,7 @@ export declare const workersAiCatalog: {
|
|
|
308
346
|
readonly description: "Quantized (int4) generative text model with 8 billion parameters from Meta.\n";
|
|
309
347
|
readonly tags: readonly [];
|
|
310
348
|
readonly properties: {
|
|
349
|
+
readonly context_window: 8192;
|
|
311
350
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
|
|
312
351
|
};
|
|
313
352
|
}, {
|
|
@@ -318,6 +357,7 @@ export declare const workersAiCatalog: {
|
|
|
318
357
|
readonly tags: readonly [];
|
|
319
358
|
readonly properties: {
|
|
320
359
|
readonly beta: true;
|
|
360
|
+
readonly context_window: 4096;
|
|
321
361
|
readonly info: "https://huggingface.co/TheBloke/zephyr-7B-beta-AWQ";
|
|
322
362
|
};
|
|
323
363
|
}, {
|
|
@@ -328,6 +368,7 @@ export declare const workersAiCatalog: {
|
|
|
328
368
|
readonly tags: readonly [];
|
|
329
369
|
readonly properties: {
|
|
330
370
|
readonly beta: true;
|
|
371
|
+
readonly context_window: 3500;
|
|
331
372
|
readonly lora: true;
|
|
332
373
|
};
|
|
333
374
|
}, {
|
|
@@ -338,6 +379,7 @@ export declare const workersAiCatalog: {
|
|
|
338
379
|
readonly tags: readonly [];
|
|
339
380
|
readonly properties: {
|
|
340
381
|
readonly beta: true;
|
|
382
|
+
readonly context_window: 32000;
|
|
341
383
|
readonly info: "https://huggingface.co/qwen/qwen1.5-1.8b-chat";
|
|
342
384
|
};
|
|
343
385
|
}, {
|
|
@@ -347,6 +389,7 @@ export declare const workersAiCatalog: {
|
|
|
347
389
|
readonly description: "Quantized (int4) generative text model with 8 billion parameters from Meta.";
|
|
348
390
|
readonly tags: readonly [];
|
|
349
391
|
readonly properties: {
|
|
392
|
+
readonly context_window: 8192;
|
|
350
393
|
readonly info: "https://llama.meta.com";
|
|
351
394
|
readonly terms: "https://llama.meta.com/llama3/license/#";
|
|
352
395
|
};
|
|
@@ -367,6 +410,7 @@ export declare const workersAiCatalog: {
|
|
|
367
410
|
readonly tags: readonly [];
|
|
368
411
|
readonly properties: {
|
|
369
412
|
readonly beta: true;
|
|
413
|
+
readonly context_window: 10000;
|
|
370
414
|
readonly info: "https://huggingface.co/defog/sqlcoder-7b-2";
|
|
371
415
|
readonly terms: "https://creativecommons.org/licenses/by-sa/4.0/deed.en";
|
|
372
416
|
};
|
|
@@ -378,6 +422,7 @@ export declare const workersAiCatalog: {
|
|
|
378
422
|
readonly tags: readonly [];
|
|
379
423
|
readonly properties: {
|
|
380
424
|
readonly beta: true;
|
|
425
|
+
readonly context_window: 2048;
|
|
381
426
|
readonly info: "https://huggingface.co/microsoft/phi-2";
|
|
382
427
|
};
|
|
383
428
|
}, {
|
|
@@ -386,7 +431,9 @@ export declare const workersAiCatalog: {
|
|
|
386
431
|
readonly name: "@hf/meta-llama/meta-llama-3-8b-instruct";
|
|
387
432
|
readonly description: "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.\t";
|
|
388
433
|
readonly tags: readonly [];
|
|
389
|
-
readonly properties: {
|
|
434
|
+
readonly properties: {
|
|
435
|
+
readonly context_window: 8192;
|
|
436
|
+
};
|
|
390
437
|
}, {
|
|
391
438
|
readonly id: "0f002249-7d86-4698-aabf-8529ed86cefb";
|
|
392
439
|
readonly source: 2;
|
|
@@ -395,11 +442,9 @@ export declare const workersAiCatalog: {
|
|
|
395
442
|
readonly tags: readonly [];
|
|
396
443
|
readonly properties: {
|
|
397
444
|
readonly beta: true;
|
|
445
|
+
readonly context_window: 8192;
|
|
398
446
|
readonly info: "https://ai.google.dev/gemma/docs";
|
|
399
447
|
readonly lora: true;
|
|
400
|
-
readonly max_batch_prefill_tokens: 2048;
|
|
401
|
-
readonly max_input_length: 1512;
|
|
402
|
-
readonly max_total_tokens: 2048;
|
|
403
448
|
readonly terms: "https://ai.google.dev/gemma/terms";
|
|
404
449
|
};
|
|
405
450
|
}, {
|
|
@@ -410,6 +455,7 @@ export declare const workersAiCatalog: {
|
|
|
410
455
|
readonly tags: readonly [];
|
|
411
456
|
readonly properties: {
|
|
412
457
|
readonly beta: true;
|
|
458
|
+
readonly context_window: 7500;
|
|
413
459
|
readonly info: "https://huggingface.co/qwen/qwen1.5-14b-chat-awq";
|
|
414
460
|
};
|
|
415
461
|
}, {
|
|
@@ -420,6 +466,7 @@ export declare const workersAiCatalog: {
|
|
|
420
466
|
readonly tags: readonly [];
|
|
421
467
|
readonly properties: {
|
|
422
468
|
readonly beta: true;
|
|
469
|
+
readonly context_window: 8192;
|
|
423
470
|
readonly info: "https://huggingface.co/openchat/openchat-3.5-0106";
|
|
424
471
|
};
|
|
425
472
|
}];
|
|
@@ -12,6 +12,7 @@ export const workersAiCatalog = {
|
|
|
12
12
|
tags: [],
|
|
13
13
|
properties: {
|
|
14
14
|
beta: true,
|
|
15
|
+
context_window: 32000,
|
|
15
16
|
info: 'https://huggingface.co/qwen/qwen1.5-0.5b-chat',
|
|
16
17
|
},
|
|
17
18
|
},
|
|
@@ -23,6 +24,7 @@ export const workersAiCatalog = {
|
|
|
23
24
|
tags: [],
|
|
24
25
|
properties: {
|
|
25
26
|
beta: true,
|
|
27
|
+
context_window: 8192,
|
|
26
28
|
lora: true,
|
|
27
29
|
},
|
|
28
30
|
},
|
|
@@ -34,6 +36,7 @@ export const workersAiCatalog = {
|
|
|
34
36
|
tags: [],
|
|
35
37
|
properties: {
|
|
36
38
|
beta: true,
|
|
39
|
+
context_window: 4096,
|
|
37
40
|
info: 'https://huggingface.co/Nexusflow/Starling-LM-7B-beta',
|
|
38
41
|
max_batch_prefill_tokens: 8192,
|
|
39
42
|
max_input_length: 3072,
|
|
@@ -47,6 +50,7 @@ export const workersAiCatalog = {
|
|
|
47
50
|
description: 'Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.',
|
|
48
51
|
tags: [],
|
|
49
52
|
properties: {
|
|
53
|
+
context_window: 7968,
|
|
50
54
|
info: 'https://llama.meta.com',
|
|
51
55
|
terms: 'https://llama.meta.com/llama3/license/#',
|
|
52
56
|
},
|
|
@@ -58,6 +62,7 @@ export const workersAiCatalog = {
|
|
|
58
62
|
description: 'The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.',
|
|
59
63
|
tags: [],
|
|
60
64
|
properties: {
|
|
65
|
+
context_window: 128000,
|
|
61
66
|
terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE',
|
|
62
67
|
},
|
|
63
68
|
},
|
|
@@ -69,6 +74,7 @@ export const workersAiCatalog = {
|
|
|
69
74
|
tags: [],
|
|
70
75
|
properties: {
|
|
71
76
|
beta: true,
|
|
77
|
+
context_window: 4096,
|
|
72
78
|
},
|
|
73
79
|
},
|
|
74
80
|
{
|
|
@@ -79,8 +85,17 @@ export const workersAiCatalog = {
|
|
|
79
85
|
tags: [],
|
|
80
86
|
properties: {
|
|
81
87
|
beta: true,
|
|
88
|
+
context_window: 4096,
|
|
82
89
|
},
|
|
83
90
|
},
|
|
91
|
+
{
|
|
92
|
+
id: 'cc80437b-9a8d-4f1a-9c77-9aaf0d226922',
|
|
93
|
+
source: 1,
|
|
94
|
+
name: '@cf/meta/llama-guard-3-8b',
|
|
95
|
+
description: 'Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.',
|
|
96
|
+
tags: [],
|
|
97
|
+
properties: {},
|
|
98
|
+
},
|
|
84
99
|
{
|
|
85
100
|
id: 'ca54bcd6-0d98-4739-9b3b-5c8b4402193d',
|
|
86
101
|
source: 1,
|
|
@@ -89,6 +104,7 @@ export const workersAiCatalog = {
|
|
|
89
104
|
tags: [],
|
|
90
105
|
properties: {
|
|
91
106
|
beta: true,
|
|
107
|
+
context_window: 4096,
|
|
92
108
|
info: 'https://ai.meta.com/llama/',
|
|
93
109
|
terms: 'https://ai.meta.com/resources/models-and-libraries/llama-downloads/',
|
|
94
110
|
},
|
|
@@ -101,6 +117,7 @@ export const workersAiCatalog = {
|
|
|
101
117
|
tags: [],
|
|
102
118
|
properties: {
|
|
103
119
|
beta: true,
|
|
120
|
+
context_window: 2824,
|
|
104
121
|
info: 'https://mistral.ai/news/announcing-mistral-7b/',
|
|
105
122
|
lora: true,
|
|
106
123
|
},
|
|
@@ -113,6 +130,7 @@ export const workersAiCatalog = {
|
|
|
113
130
|
tags: [],
|
|
114
131
|
properties: {
|
|
115
132
|
beta: true,
|
|
133
|
+
context_window: 15000,
|
|
116
134
|
lora: true,
|
|
117
135
|
},
|
|
118
136
|
},
|
|
@@ -124,6 +142,7 @@ export const workersAiCatalog = {
|
|
|
124
142
|
tags: [],
|
|
125
143
|
properties: {
|
|
126
144
|
beta: true,
|
|
145
|
+
context_window: 2048,
|
|
127
146
|
info: 'https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0',
|
|
128
147
|
},
|
|
129
148
|
},
|
|
@@ -135,6 +154,7 @@ export const workersAiCatalog = {
|
|
|
135
154
|
tags: [],
|
|
136
155
|
properties: {
|
|
137
156
|
beta: true,
|
|
157
|
+
context_window: 3072,
|
|
138
158
|
info: 'https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2',
|
|
139
159
|
lora: true,
|
|
140
160
|
max_batch_prefill_tokens: 8192,
|
|
@@ -150,6 +170,7 @@ export const workersAiCatalog = {
|
|
|
150
170
|
tags: [],
|
|
151
171
|
properties: {
|
|
152
172
|
beta: true,
|
|
173
|
+
context_window: 15000,
|
|
153
174
|
},
|
|
154
175
|
},
|
|
155
176
|
{
|
|
@@ -159,6 +180,7 @@ export const workersAiCatalog = {
|
|
|
159
180
|
description: 'DeepSeek-R1-Distill-Qwen-32B is a model distilled from DeepSeek-R1 based on Qwen2.5. It outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.',
|
|
160
181
|
tags: [],
|
|
161
182
|
properties: {
|
|
183
|
+
context_window: 80000,
|
|
162
184
|
terms: 'https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE',
|
|
163
185
|
},
|
|
164
186
|
},
|
|
@@ -170,6 +192,7 @@ export const workersAiCatalog = {
|
|
|
170
192
|
tags: [],
|
|
171
193
|
properties: {
|
|
172
194
|
beta: true,
|
|
195
|
+
context_window: 4096,
|
|
173
196
|
info: 'https://huggingface.co/TheBloke/DiscoLM_German_7b_v1-AWQ',
|
|
174
197
|
},
|
|
175
198
|
},
|
|
@@ -179,7 +202,9 @@ export const workersAiCatalog = {
|
|
|
179
202
|
name: '@cf/meta/llama-2-7b-chat-int8',
|
|
180
203
|
description: 'Quantized (int8) generative text model with 7 billion parameters from Meta',
|
|
181
204
|
tags: [],
|
|
182
|
-
properties: {
|
|
205
|
+
properties: {
|
|
206
|
+
context_window: 8192,
|
|
207
|
+
},
|
|
183
208
|
},
|
|
184
209
|
{
|
|
185
210
|
id: '9b9c87c6-d4b7-494c-b177-87feab5904db',
|
|
@@ -188,6 +213,7 @@ export const workersAiCatalog = {
|
|
|
188
213
|
description: 'Llama 3.1 8B quantized to FP8 precision',
|
|
189
214
|
tags: [],
|
|
190
215
|
properties: {
|
|
216
|
+
context_window: 32000,
|
|
191
217
|
terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE',
|
|
192
218
|
},
|
|
193
219
|
},
|
|
@@ -199,6 +225,7 @@ export const workersAiCatalog = {
|
|
|
199
225
|
tags: [],
|
|
200
226
|
properties: {
|
|
201
227
|
beta: true,
|
|
228
|
+
context_window: 4096,
|
|
202
229
|
info: 'https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-AWQ',
|
|
203
230
|
},
|
|
204
231
|
},
|
|
@@ -210,6 +237,7 @@ export const workersAiCatalog = {
|
|
|
210
237
|
tags: [],
|
|
211
238
|
properties: {
|
|
212
239
|
beta: true,
|
|
240
|
+
context_window: 20000,
|
|
213
241
|
info: 'https://huggingface.co/qwen/qwen1.5-7b-chat-awq',
|
|
214
242
|
},
|
|
215
243
|
},
|
|
@@ -220,6 +248,7 @@ export const workersAiCatalog = {
|
|
|
220
248
|
description: 'The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.',
|
|
221
249
|
tags: [],
|
|
222
250
|
properties: {
|
|
251
|
+
context_window: 128000,
|
|
223
252
|
terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE',
|
|
224
253
|
},
|
|
225
254
|
},
|
|
@@ -231,6 +260,7 @@ export const workersAiCatalog = {
|
|
|
231
260
|
tags: [],
|
|
232
261
|
properties: {
|
|
233
262
|
beta: true,
|
|
263
|
+
context_window: 4096,
|
|
234
264
|
info: 'https://huggingface.co/TheBloke/Llama-2-13B-chat-AWQ',
|
|
235
265
|
},
|
|
236
266
|
},
|
|
@@ -242,6 +272,7 @@ export const workersAiCatalog = {
|
|
|
242
272
|
tags: [],
|
|
243
273
|
properties: {
|
|
244
274
|
beta: true,
|
|
275
|
+
context_window: 4096,
|
|
245
276
|
terms: 'https://huggingface.co/TheBloke/deepseek-coder-6.7B-base-AWQ',
|
|
246
277
|
},
|
|
247
278
|
},
|
|
@@ -253,6 +284,7 @@ export const workersAiCatalog = {
|
|
|
253
284
|
tags: [],
|
|
254
285
|
properties: {
|
|
255
286
|
beta: true,
|
|
287
|
+
context_window: 8192,
|
|
256
288
|
lora: true,
|
|
257
289
|
},
|
|
258
290
|
},
|
|
@@ -263,6 +295,7 @@ export const workersAiCatalog = {
|
|
|
263
295
|
description: 'Llama 3.3 70B quantized to fp8 precision, optimized to be faster.',
|
|
264
296
|
tags: [],
|
|
265
297
|
properties: {
|
|
298
|
+
context_window: 24000,
|
|
266
299
|
terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_3/LICENSE',
|
|
267
300
|
},
|
|
268
301
|
},
|
|
@@ -274,6 +307,7 @@ export const workersAiCatalog = {
|
|
|
274
307
|
tags: [],
|
|
275
308
|
properties: {
|
|
276
309
|
beta: true,
|
|
310
|
+
context_window: 4096,
|
|
277
311
|
},
|
|
278
312
|
},
|
|
279
313
|
{
|
|
@@ -284,6 +318,7 @@ export const workersAiCatalog = {
|
|
|
284
318
|
tags: [],
|
|
285
319
|
properties: {
|
|
286
320
|
beta: true,
|
|
321
|
+
context_window: 4096,
|
|
287
322
|
terms: 'https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-AWQ',
|
|
288
323
|
},
|
|
289
324
|
},
|
|
@@ -295,6 +330,7 @@ export const workersAiCatalog = {
|
|
|
295
330
|
tags: [],
|
|
296
331
|
properties: {
|
|
297
332
|
beta: true,
|
|
333
|
+
context_window: 4096,
|
|
298
334
|
info: 'https://huggingface.co/deepseek-ai/deepseek-math-7b-instruct',
|
|
299
335
|
terms: 'https://github.com/deepseek-ai/DeepSeek-Math/blob/main/LICENSE-MODEL',
|
|
300
336
|
},
|
|
@@ -307,6 +343,7 @@ export const workersAiCatalog = {
|
|
|
307
343
|
tags: [],
|
|
308
344
|
properties: {
|
|
309
345
|
beta: true,
|
|
346
|
+
context_window: 4096,
|
|
310
347
|
info: 'https://huggingface.co/tiiuae/falcon-7b-instruct',
|
|
311
348
|
},
|
|
312
349
|
},
|
|
@@ -318,6 +355,7 @@ export const workersAiCatalog = {
|
|
|
318
355
|
tags: [],
|
|
319
356
|
properties: {
|
|
320
357
|
beta: true,
|
|
358
|
+
context_window: 24000,
|
|
321
359
|
function_calling: true,
|
|
322
360
|
info: 'https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B',
|
|
323
361
|
},
|
|
@@ -329,6 +367,7 @@ export const workersAiCatalog = {
|
|
|
329
367
|
description: 'The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.',
|
|
330
368
|
tags: [],
|
|
331
369
|
properties: {
|
|
370
|
+
context_window: 7968,
|
|
332
371
|
terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE',
|
|
333
372
|
},
|
|
334
373
|
},
|
|
@@ -339,6 +378,7 @@ export const workersAiCatalog = {
|
|
|
339
378
|
description: 'Quantized (int4) generative text model with 8 billion parameters from Meta.\n',
|
|
340
379
|
tags: [],
|
|
341
380
|
properties: {
|
|
381
|
+
context_window: 8192,
|
|
342
382
|
terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE',
|
|
343
383
|
},
|
|
344
384
|
},
|
|
@@ -350,6 +390,7 @@ export const workersAiCatalog = {
|
|
|
350
390
|
tags: [],
|
|
351
391
|
properties: {
|
|
352
392
|
beta: true,
|
|
393
|
+
context_window: 4096,
|
|
353
394
|
info: 'https://huggingface.co/TheBloke/zephyr-7B-beta-AWQ',
|
|
354
395
|
},
|
|
355
396
|
},
|
|
@@ -361,6 +402,7 @@ export const workersAiCatalog = {
|
|
|
361
402
|
tags: [],
|
|
362
403
|
properties: {
|
|
363
404
|
beta: true,
|
|
405
|
+
context_window: 3500,
|
|
364
406
|
lora: true,
|
|
365
407
|
},
|
|
366
408
|
},
|
|
@@ -372,6 +414,7 @@ export const workersAiCatalog = {
|
|
|
372
414
|
tags: [],
|
|
373
415
|
properties: {
|
|
374
416
|
beta: true,
|
|
417
|
+
context_window: 32000,
|
|
375
418
|
info: 'https://huggingface.co/qwen/qwen1.5-1.8b-chat',
|
|
376
419
|
},
|
|
377
420
|
},
|
|
@@ -382,6 +425,7 @@ export const workersAiCatalog = {
|
|
|
382
425
|
description: 'Quantized (int4) generative text model with 8 billion parameters from Meta.',
|
|
383
426
|
tags: [],
|
|
384
427
|
properties: {
|
|
428
|
+
context_window: 8192,
|
|
385
429
|
info: 'https://llama.meta.com',
|
|
386
430
|
terms: 'https://llama.meta.com/llama3/license/#',
|
|
387
431
|
},
|
|
@@ -404,6 +448,7 @@ export const workersAiCatalog = {
|
|
|
404
448
|
tags: [],
|
|
405
449
|
properties: {
|
|
406
450
|
beta: true,
|
|
451
|
+
context_window: 10000,
|
|
407
452
|
info: 'https://huggingface.co/defog/sqlcoder-7b-2',
|
|
408
453
|
terms: 'https://creativecommons.org/licenses/by-sa/4.0/deed.en',
|
|
409
454
|
},
|
|
@@ -416,6 +461,7 @@ export const workersAiCatalog = {
|
|
|
416
461
|
tags: [],
|
|
417
462
|
properties: {
|
|
418
463
|
beta: true,
|
|
464
|
+
context_window: 2048,
|
|
419
465
|
info: 'https://huggingface.co/microsoft/phi-2',
|
|
420
466
|
},
|
|
421
467
|
},
|
|
@@ -425,7 +471,9 @@ export const workersAiCatalog = {
|
|
|
425
471
|
name: '@hf/meta-llama/meta-llama-3-8b-instruct',
|
|
426
472
|
description: 'Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.\t',
|
|
427
473
|
tags: [],
|
|
428
|
-
properties: {
|
|
474
|
+
properties: {
|
|
475
|
+
context_window: 8192,
|
|
476
|
+
},
|
|
429
477
|
},
|
|
430
478
|
{
|
|
431
479
|
id: '0f002249-7d86-4698-aabf-8529ed86cefb',
|
|
@@ -435,11 +483,9 @@ export const workersAiCatalog = {
|
|
|
435
483
|
tags: [],
|
|
436
484
|
properties: {
|
|
437
485
|
beta: true,
|
|
486
|
+
context_window: 8192,
|
|
438
487
|
info: 'https://ai.google.dev/gemma/docs',
|
|
439
488
|
lora: true,
|
|
440
|
-
max_batch_prefill_tokens: 2048,
|
|
441
|
-
max_input_length: 1512,
|
|
442
|
-
max_total_tokens: 2048,
|
|
443
489
|
terms: 'https://ai.google.dev/gemma/terms',
|
|
444
490
|
},
|
|
445
491
|
},
|
|
@@ -451,6 +497,7 @@ export const workersAiCatalog = {
|
|
|
451
497
|
tags: [],
|
|
452
498
|
properties: {
|
|
453
499
|
beta: true,
|
|
500
|
+
context_window: 7500,
|
|
454
501
|
info: 'https://huggingface.co/qwen/qwen1.5-14b-chat-awq',
|
|
455
502
|
},
|
|
456
503
|
},
|
|
@@ -462,6 +509,7 @@ export const workersAiCatalog = {
|
|
|
462
509
|
tags: [],
|
|
463
510
|
properties: {
|
|
464
511
|
beta: true,
|
|
512
|
+
context_window: 8192,
|
|
465
513
|
info: 'https://huggingface.co/openchat/openchat-3.5-0106',
|
|
466
514
|
},
|
|
467
515
|
},
|
package/dist/super-ai/index.d.ts
CHANGED
|
@@ -50,6 +50,7 @@ export declare const possibilities_mc_generic: readonly [...{
|
|
|
50
50
|
readonly tags: readonly [];
|
|
51
51
|
readonly properties: {
|
|
52
52
|
readonly beta: true;
|
|
53
|
+
readonly context_window: 24000;
|
|
53
54
|
readonly function_calling: true;
|
|
54
55
|
readonly info: "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B";
|
|
55
56
|
};
|
|
@@ -78,6 +79,7 @@ export declare const possibilities_mc_extraction: readonly [...{
|
|
|
78
79
|
readonly tags: readonly [];
|
|
79
80
|
readonly properties: {
|
|
80
81
|
readonly beta: true;
|
|
82
|
+
readonly context_window: 24000;
|
|
81
83
|
readonly function_calling: true;
|
|
82
84
|
readonly info: "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B";
|
|
83
85
|
};
|
|
@@ -93,6 +95,7 @@ export declare const possibilities_mc_tagging: readonly [...{
|
|
|
93
95
|
readonly tags: readonly [];
|
|
94
96
|
readonly properties: {
|
|
95
97
|
readonly beta: true;
|
|
98
|
+
readonly context_window: 24000;
|
|
96
99
|
readonly function_calling: true;
|
|
97
100
|
readonly info: "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B";
|
|
98
101
|
};
|
|
@@ -108,6 +111,7 @@ export declare const possibilities_mc_sentiment: readonly [...{
|
|
|
108
111
|
readonly tags: readonly [];
|
|
109
112
|
readonly properties: {
|
|
110
113
|
readonly beta: true;
|
|
114
|
+
readonly context_window: 24000;
|
|
111
115
|
readonly function_calling: true;
|
|
112
116
|
readonly info: "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B";
|
|
113
117
|
};
|
|
@@ -123,6 +127,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
123
127
|
readonly tags: readonly [];
|
|
124
128
|
readonly properties: {
|
|
125
129
|
readonly beta: true;
|
|
130
|
+
readonly context_window: 32000;
|
|
126
131
|
readonly info: "https://huggingface.co/qwen/qwen1.5-0.5b-chat";
|
|
127
132
|
};
|
|
128
133
|
}, {
|
|
@@ -133,6 +138,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
133
138
|
readonly tags: readonly [];
|
|
134
139
|
readonly properties: {
|
|
135
140
|
readonly beta: true;
|
|
141
|
+
readonly context_window: 8192;
|
|
136
142
|
readonly lora: true;
|
|
137
143
|
};
|
|
138
144
|
}, {
|
|
@@ -143,6 +149,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
143
149
|
readonly tags: readonly [];
|
|
144
150
|
readonly properties: {
|
|
145
151
|
readonly beta: true;
|
|
152
|
+
readonly context_window: 4096;
|
|
146
153
|
readonly info: "https://huggingface.co/Nexusflow/Starling-LM-7B-beta";
|
|
147
154
|
readonly max_batch_prefill_tokens: 8192;
|
|
148
155
|
readonly max_input_length: 3072;
|
|
@@ -155,6 +162,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
155
162
|
readonly description: "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.";
|
|
156
163
|
readonly tags: readonly [];
|
|
157
164
|
readonly properties: {
|
|
165
|
+
readonly context_window: 7968;
|
|
158
166
|
readonly info: "https://llama.meta.com";
|
|
159
167
|
readonly terms: "https://llama.meta.com/llama3/license/#";
|
|
160
168
|
};
|
|
@@ -165,6 +173,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
165
173
|
readonly description: "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
|
|
166
174
|
readonly tags: readonly [];
|
|
167
175
|
readonly properties: {
|
|
176
|
+
readonly context_window: 128000;
|
|
168
177
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
|
|
169
178
|
};
|
|
170
179
|
}, {
|
|
@@ -175,6 +184,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
175
184
|
readonly tags: readonly [];
|
|
176
185
|
readonly properties: {
|
|
177
186
|
readonly beta: true;
|
|
187
|
+
readonly context_window: 4096;
|
|
178
188
|
};
|
|
179
189
|
}, {
|
|
180
190
|
readonly id: "d2ba5c6b-bbb7-49d6-b466-900654870cd6";
|
|
@@ -184,7 +194,15 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
184
194
|
readonly tags: readonly [];
|
|
185
195
|
readonly properties: {
|
|
186
196
|
readonly beta: true;
|
|
197
|
+
readonly context_window: 4096;
|
|
187
198
|
};
|
|
199
|
+
}, {
|
|
200
|
+
readonly id: "cc80437b-9a8d-4f1a-9c77-9aaf0d226922";
|
|
201
|
+
readonly source: 1;
|
|
202
|
+
readonly name: "@cf/meta/llama-guard-3-8b";
|
|
203
|
+
readonly description: "Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.";
|
|
204
|
+
readonly tags: readonly [];
|
|
205
|
+
readonly properties: {};
|
|
188
206
|
}, {
|
|
189
207
|
readonly id: "ca54bcd6-0d98-4739-9b3b-5c8b4402193d";
|
|
190
208
|
readonly source: 1;
|
|
@@ -193,6 +211,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
193
211
|
readonly tags: readonly [];
|
|
194
212
|
readonly properties: {
|
|
195
213
|
readonly beta: true;
|
|
214
|
+
readonly context_window: 4096;
|
|
196
215
|
readonly info: "https://ai.meta.com/llama/";
|
|
197
216
|
readonly terms: "https://ai.meta.com/resources/models-and-libraries/llama-downloads/";
|
|
198
217
|
};
|
|
@@ -204,6 +223,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
204
223
|
readonly tags: readonly [];
|
|
205
224
|
readonly properties: {
|
|
206
225
|
readonly beta: true;
|
|
226
|
+
readonly context_window: 2824;
|
|
207
227
|
readonly info: "https://mistral.ai/news/announcing-mistral-7b/";
|
|
208
228
|
readonly lora: true;
|
|
209
229
|
};
|
|
@@ -215,6 +235,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
215
235
|
readonly tags: readonly [];
|
|
216
236
|
readonly properties: {
|
|
217
237
|
readonly beta: true;
|
|
238
|
+
readonly context_window: 15000;
|
|
218
239
|
readonly lora: true;
|
|
219
240
|
};
|
|
220
241
|
}, {
|
|
@@ -225,6 +246,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
225
246
|
readonly tags: readonly [];
|
|
226
247
|
readonly properties: {
|
|
227
248
|
readonly beta: true;
|
|
249
|
+
readonly context_window: 2048;
|
|
228
250
|
readonly info: "https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0";
|
|
229
251
|
};
|
|
230
252
|
}, {
|
|
@@ -235,6 +257,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
235
257
|
readonly tags: readonly [];
|
|
236
258
|
readonly properties: {
|
|
237
259
|
readonly beta: true;
|
|
260
|
+
readonly context_window: 3072;
|
|
238
261
|
readonly info: "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2";
|
|
239
262
|
readonly lora: true;
|
|
240
263
|
readonly max_batch_prefill_tokens: 8192;
|
|
@@ -249,6 +272,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
249
272
|
readonly tags: readonly [];
|
|
250
273
|
readonly properties: {
|
|
251
274
|
readonly beta: true;
|
|
275
|
+
readonly context_window: 15000;
|
|
252
276
|
};
|
|
253
277
|
}, {
|
|
254
278
|
readonly id: "ad01ab83-baf8-4e7b-8fed-a0a219d4eb45";
|
|
@@ -257,6 +281,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
257
281
|
readonly description: "DeepSeek-R1-Distill-Qwen-32B is a model distilled from DeepSeek-R1 based on Qwen2.5. It outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.";
|
|
258
282
|
readonly tags: readonly [];
|
|
259
283
|
readonly properties: {
|
|
284
|
+
readonly context_window: 80000;
|
|
260
285
|
readonly terms: "https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE";
|
|
261
286
|
};
|
|
262
287
|
}, {
|
|
@@ -267,6 +292,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
267
292
|
readonly tags: readonly [];
|
|
268
293
|
readonly properties: {
|
|
269
294
|
readonly beta: true;
|
|
295
|
+
readonly context_window: 4096;
|
|
270
296
|
readonly info: "https://huggingface.co/TheBloke/DiscoLM_German_7b_v1-AWQ";
|
|
271
297
|
};
|
|
272
298
|
}, {
|
|
@@ -275,7 +301,9 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
275
301
|
readonly name: "@cf/meta/llama-2-7b-chat-int8";
|
|
276
302
|
readonly description: "Quantized (int8) generative text model with 7 billion parameters from Meta";
|
|
277
303
|
readonly tags: readonly [];
|
|
278
|
-
readonly properties: {
|
|
304
|
+
readonly properties: {
|
|
305
|
+
readonly context_window: 8192;
|
|
306
|
+
};
|
|
279
307
|
}, {
|
|
280
308
|
readonly id: "9b9c87c6-d4b7-494c-b177-87feab5904db";
|
|
281
309
|
readonly source: 1;
|
|
@@ -283,6 +311,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
283
311
|
readonly description: "Llama 3.1 8B quantized to FP8 precision";
|
|
284
312
|
readonly tags: readonly [];
|
|
285
313
|
readonly properties: {
|
|
314
|
+
readonly context_window: 32000;
|
|
286
315
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
|
|
287
316
|
};
|
|
288
317
|
}, {
|
|
@@ -293,6 +322,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
293
322
|
readonly tags: readonly [];
|
|
294
323
|
readonly properties: {
|
|
295
324
|
readonly beta: true;
|
|
325
|
+
readonly context_window: 4096;
|
|
296
326
|
readonly info: "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-AWQ";
|
|
297
327
|
};
|
|
298
328
|
}, {
|
|
@@ -303,6 +333,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
303
333
|
readonly tags: readonly [];
|
|
304
334
|
readonly properties: {
|
|
305
335
|
readonly beta: true;
|
|
336
|
+
readonly context_window: 20000;
|
|
306
337
|
readonly info: "https://huggingface.co/qwen/qwen1.5-7b-chat-awq";
|
|
307
338
|
};
|
|
308
339
|
}, {
|
|
@@ -312,6 +343,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
312
343
|
readonly description: "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
|
|
313
344
|
readonly tags: readonly [];
|
|
314
345
|
readonly properties: {
|
|
346
|
+
readonly context_window: 128000;
|
|
315
347
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
|
|
316
348
|
};
|
|
317
349
|
}, {
|
|
@@ -322,6 +354,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
322
354
|
readonly tags: readonly [];
|
|
323
355
|
readonly properties: {
|
|
324
356
|
readonly beta: true;
|
|
357
|
+
readonly context_window: 4096;
|
|
325
358
|
readonly info: "https://huggingface.co/TheBloke/Llama-2-13B-chat-AWQ";
|
|
326
359
|
};
|
|
327
360
|
}, {
|
|
@@ -332,6 +365,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
332
365
|
readonly tags: readonly [];
|
|
333
366
|
readonly properties: {
|
|
334
367
|
readonly beta: true;
|
|
368
|
+
readonly context_window: 4096;
|
|
335
369
|
readonly terms: "https://huggingface.co/TheBloke/deepseek-coder-6.7B-base-AWQ";
|
|
336
370
|
};
|
|
337
371
|
}, {
|
|
@@ -342,6 +376,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
342
376
|
readonly tags: readonly [];
|
|
343
377
|
readonly properties: {
|
|
344
378
|
readonly beta: true;
|
|
379
|
+
readonly context_window: 8192;
|
|
345
380
|
readonly lora: true;
|
|
346
381
|
};
|
|
347
382
|
}, {
|
|
@@ -351,6 +386,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
351
386
|
readonly description: "Llama 3.3 70B quantized to fp8 precision, optimized to be faster.";
|
|
352
387
|
readonly tags: readonly [];
|
|
353
388
|
readonly properties: {
|
|
389
|
+
readonly context_window: 24000;
|
|
354
390
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_3/LICENSE";
|
|
355
391
|
};
|
|
356
392
|
}, {
|
|
@@ -361,6 +397,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
361
397
|
readonly tags: readonly [];
|
|
362
398
|
readonly properties: {
|
|
363
399
|
readonly beta: true;
|
|
400
|
+
readonly context_window: 4096;
|
|
364
401
|
};
|
|
365
402
|
}, {
|
|
366
403
|
readonly id: "60474554-f03b-4ff4-8ecc-c1b7c71d7b29";
|
|
@@ -370,6 +407,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
370
407
|
readonly tags: readonly [];
|
|
371
408
|
readonly properties: {
|
|
372
409
|
readonly beta: true;
|
|
410
|
+
readonly context_window: 4096;
|
|
373
411
|
readonly terms: "https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-AWQ";
|
|
374
412
|
};
|
|
375
413
|
}, {
|
|
@@ -380,6 +418,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
380
418
|
readonly tags: readonly [];
|
|
381
419
|
readonly properties: {
|
|
382
420
|
readonly beta: true;
|
|
421
|
+
readonly context_window: 4096;
|
|
383
422
|
readonly info: "https://huggingface.co/deepseek-ai/deepseek-math-7b-instruct";
|
|
384
423
|
readonly terms: "https://github.com/deepseek-ai/DeepSeek-Math/blob/main/LICENSE-MODEL";
|
|
385
424
|
};
|
|
@@ -391,6 +430,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
391
430
|
readonly tags: readonly [];
|
|
392
431
|
readonly properties: {
|
|
393
432
|
readonly beta: true;
|
|
433
|
+
readonly context_window: 4096;
|
|
394
434
|
readonly info: "https://huggingface.co/tiiuae/falcon-7b-instruct";
|
|
395
435
|
};
|
|
396
436
|
}, {
|
|
@@ -401,6 +441,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
401
441
|
readonly tags: readonly [];
|
|
402
442
|
readonly properties: {
|
|
403
443
|
readonly beta: true;
|
|
444
|
+
readonly context_window: 24000;
|
|
404
445
|
readonly function_calling: true;
|
|
405
446
|
readonly info: "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B";
|
|
406
447
|
};
|
|
@@ -411,6 +452,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
411
452
|
readonly description: "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.";
|
|
412
453
|
readonly tags: readonly [];
|
|
413
454
|
readonly properties: {
|
|
455
|
+
readonly context_window: 7968;
|
|
414
456
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
|
|
415
457
|
};
|
|
416
458
|
}, {
|
|
@@ -420,6 +462,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
420
462
|
readonly description: "Quantized (int4) generative text model with 8 billion parameters from Meta.\n";
|
|
421
463
|
readonly tags: readonly [];
|
|
422
464
|
readonly properties: {
|
|
465
|
+
readonly context_window: 8192;
|
|
423
466
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
|
|
424
467
|
};
|
|
425
468
|
}, {
|
|
@@ -430,6 +473,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
430
473
|
readonly tags: readonly [];
|
|
431
474
|
readonly properties: {
|
|
432
475
|
readonly beta: true;
|
|
476
|
+
readonly context_window: 4096;
|
|
433
477
|
readonly info: "https://huggingface.co/TheBloke/zephyr-7B-beta-AWQ";
|
|
434
478
|
};
|
|
435
479
|
}, {
|
|
@@ -440,6 +484,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
440
484
|
readonly tags: readonly [];
|
|
441
485
|
readonly properties: {
|
|
442
486
|
readonly beta: true;
|
|
487
|
+
readonly context_window: 3500;
|
|
443
488
|
readonly lora: true;
|
|
444
489
|
};
|
|
445
490
|
}, {
|
|
@@ -450,6 +495,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
450
495
|
readonly tags: readonly [];
|
|
451
496
|
readonly properties: {
|
|
452
497
|
readonly beta: true;
|
|
498
|
+
readonly context_window: 32000;
|
|
453
499
|
readonly info: "https://huggingface.co/qwen/qwen1.5-1.8b-chat";
|
|
454
500
|
};
|
|
455
501
|
}, {
|
|
@@ -459,6 +505,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
459
505
|
readonly description: "Quantized (int4) generative text model with 8 billion parameters from Meta.";
|
|
460
506
|
readonly tags: readonly [];
|
|
461
507
|
readonly properties: {
|
|
508
|
+
readonly context_window: 8192;
|
|
462
509
|
readonly info: "https://llama.meta.com";
|
|
463
510
|
readonly terms: "https://llama.meta.com/llama3/license/#";
|
|
464
511
|
};
|
|
@@ -479,6 +526,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
479
526
|
readonly tags: readonly [];
|
|
480
527
|
readonly properties: {
|
|
481
528
|
readonly beta: true;
|
|
529
|
+
readonly context_window: 10000;
|
|
482
530
|
readonly info: "https://huggingface.co/defog/sqlcoder-7b-2";
|
|
483
531
|
readonly terms: "https://creativecommons.org/licenses/by-sa/4.0/deed.en";
|
|
484
532
|
};
|
|
@@ -490,6 +538,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
490
538
|
readonly tags: readonly [];
|
|
491
539
|
readonly properties: {
|
|
492
540
|
readonly beta: true;
|
|
541
|
+
readonly context_window: 2048;
|
|
493
542
|
readonly info: "https://huggingface.co/microsoft/phi-2";
|
|
494
543
|
};
|
|
495
544
|
}, {
|
|
@@ -498,7 +547,9 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
498
547
|
readonly name: "@hf/meta-llama/meta-llama-3-8b-instruct";
|
|
499
548
|
readonly description: "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.\t";
|
|
500
549
|
readonly tags: readonly [];
|
|
501
|
-
readonly properties: {
|
|
550
|
+
readonly properties: {
|
|
551
|
+
readonly context_window: 8192;
|
|
552
|
+
};
|
|
502
553
|
}, {
|
|
503
554
|
readonly id: "0f002249-7d86-4698-aabf-8529ed86cefb";
|
|
504
555
|
readonly source: 2;
|
|
@@ -507,11 +558,9 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
507
558
|
readonly tags: readonly [];
|
|
508
559
|
readonly properties: {
|
|
509
560
|
readonly beta: true;
|
|
561
|
+
readonly context_window: 8192;
|
|
510
562
|
readonly info: "https://ai.google.dev/gemma/docs";
|
|
511
563
|
readonly lora: true;
|
|
512
|
-
readonly max_batch_prefill_tokens: 2048;
|
|
513
|
-
readonly max_input_length: 1512;
|
|
514
|
-
readonly max_total_tokens: 2048;
|
|
515
564
|
readonly terms: "https://ai.google.dev/gemma/terms";
|
|
516
565
|
};
|
|
517
566
|
}, {
|
|
@@ -522,6 +571,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
522
571
|
readonly tags: readonly [];
|
|
523
572
|
readonly properties: {
|
|
524
573
|
readonly beta: true;
|
|
574
|
+
readonly context_window: 7500;
|
|
525
575
|
readonly info: "https://huggingface.co/qwen/qwen1.5-14b-chat-awq";
|
|
526
576
|
};
|
|
527
577
|
}, {
|
|
@@ -532,6 +582,7 @@ export declare const possibilities_mc_safety: readonly [...{
|
|
|
532
582
|
readonly tags: readonly [];
|
|
533
583
|
readonly properties: {
|
|
534
584
|
readonly beta: true;
|
|
585
|
+
readonly context_window: 8192;
|
|
535
586
|
readonly info: "https://huggingface.co/openchat/openchat-3.5-0106";
|
|
536
587
|
};
|
|
537
588
|
}];
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@chainfuse/types",
|
|
3
|
-
"version": "1.6.
|
|
3
|
+
"version": "1.6.9",
|
|
4
4
|
"description": "",
|
|
5
5
|
"author": "ChainFuse",
|
|
6
6
|
"homepage": "https://github.com/ChainFuse/packages/tree/main/packages/types#readme",
|
|
@@ -86,8 +86,8 @@
|
|
|
86
86
|
},
|
|
87
87
|
"prettier": "@demosjarco/prettier-config",
|
|
88
88
|
"devDependencies": {
|
|
89
|
-
"@cloudflare/workers-types": "^4.
|
|
89
|
+
"@cloudflare/workers-types": "^4.20250224.0",
|
|
90
90
|
"@types/json-schema": "^7.0.15"
|
|
91
91
|
},
|
|
92
|
-
"gitHead": "
|
|
92
|
+
"gitHead": "9d842a11f77e62dab485f3965c724827b44f688f"
|
|
93
93
|
}
|