@chainfuse/types 1.6.7 → 1.6.9

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -11,6 +11,7 @@ export declare const workersAiCatalog: {
11
11
  readonly tags: readonly [];
12
12
  readonly properties: {
13
13
  readonly beta: true;
14
+ readonly context_window: 32000;
14
15
  readonly info: "https://huggingface.co/qwen/qwen1.5-0.5b-chat";
15
16
  };
16
17
  }, {
@@ -21,6 +22,7 @@ export declare const workersAiCatalog: {
21
22
  readonly tags: readonly [];
22
23
  readonly properties: {
23
24
  readonly beta: true;
25
+ readonly context_window: 8192;
24
26
  readonly lora: true;
25
27
  };
26
28
  }, {
@@ -31,6 +33,7 @@ export declare const workersAiCatalog: {
31
33
  readonly tags: readonly [];
32
34
  readonly properties: {
33
35
  readonly beta: true;
36
+ readonly context_window: 4096;
34
37
  readonly info: "https://huggingface.co/Nexusflow/Starling-LM-7B-beta";
35
38
  readonly max_batch_prefill_tokens: 8192;
36
39
  readonly max_input_length: 3072;
@@ -43,6 +46,7 @@ export declare const workersAiCatalog: {
43
46
  readonly description: "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.";
44
47
  readonly tags: readonly [];
45
48
  readonly properties: {
49
+ readonly context_window: 7968;
46
50
  readonly info: "https://llama.meta.com";
47
51
  readonly terms: "https://llama.meta.com/llama3/license/#";
48
52
  };
@@ -53,6 +57,7 @@ export declare const workersAiCatalog: {
53
57
  readonly description: "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
54
58
  readonly tags: readonly [];
55
59
  readonly properties: {
60
+ readonly context_window: 128000;
56
61
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
57
62
  };
58
63
  }, {
@@ -63,6 +68,7 @@ export declare const workersAiCatalog: {
63
68
  readonly tags: readonly [];
64
69
  readonly properties: {
65
70
  readonly beta: true;
71
+ readonly context_window: 4096;
66
72
  };
67
73
  }, {
68
74
  readonly id: "d2ba5c6b-bbb7-49d6-b466-900654870cd6";
@@ -72,7 +78,15 @@ export declare const workersAiCatalog: {
72
78
  readonly tags: readonly [];
73
79
  readonly properties: {
74
80
  readonly beta: true;
81
+ readonly context_window: 4096;
75
82
  };
83
+ }, {
84
+ readonly id: "cc80437b-9a8d-4f1a-9c77-9aaf0d226922";
85
+ readonly source: 1;
86
+ readonly name: "@cf/meta/llama-guard-3-8b";
87
+ readonly description: "Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.";
88
+ readonly tags: readonly [];
89
+ readonly properties: {};
76
90
  }, {
77
91
  readonly id: "ca54bcd6-0d98-4739-9b3b-5c8b4402193d";
78
92
  readonly source: 1;
@@ -81,6 +95,7 @@ export declare const workersAiCatalog: {
81
95
  readonly tags: readonly [];
82
96
  readonly properties: {
83
97
  readonly beta: true;
98
+ readonly context_window: 4096;
84
99
  readonly info: "https://ai.meta.com/llama/";
85
100
  readonly terms: "https://ai.meta.com/resources/models-and-libraries/llama-downloads/";
86
101
  };
@@ -92,6 +107,7 @@ export declare const workersAiCatalog: {
92
107
  readonly tags: readonly [];
93
108
  readonly properties: {
94
109
  readonly beta: true;
110
+ readonly context_window: 2824;
95
111
  readonly info: "https://mistral.ai/news/announcing-mistral-7b/";
96
112
  readonly lora: true;
97
113
  };
@@ -103,6 +119,7 @@ export declare const workersAiCatalog: {
103
119
  readonly tags: readonly [];
104
120
  readonly properties: {
105
121
  readonly beta: true;
122
+ readonly context_window: 15000;
106
123
  readonly lora: true;
107
124
  };
108
125
  }, {
@@ -113,6 +130,7 @@ export declare const workersAiCatalog: {
113
130
  readonly tags: readonly [];
114
131
  readonly properties: {
115
132
  readonly beta: true;
133
+ readonly context_window: 2048;
116
134
  readonly info: "https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0";
117
135
  };
118
136
  }, {
@@ -123,6 +141,7 @@ export declare const workersAiCatalog: {
123
141
  readonly tags: readonly [];
124
142
  readonly properties: {
125
143
  readonly beta: true;
144
+ readonly context_window: 3072;
126
145
  readonly info: "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2";
127
146
  readonly lora: true;
128
147
  readonly max_batch_prefill_tokens: 8192;
@@ -137,6 +156,7 @@ export declare const workersAiCatalog: {
137
156
  readonly tags: readonly [];
138
157
  readonly properties: {
139
158
  readonly beta: true;
159
+ readonly context_window: 15000;
140
160
  };
141
161
  }, {
142
162
  readonly id: "ad01ab83-baf8-4e7b-8fed-a0a219d4eb45";
@@ -145,6 +165,7 @@ export declare const workersAiCatalog: {
145
165
  readonly description: "DeepSeek-R1-Distill-Qwen-32B is a model distilled from DeepSeek-R1 based on Qwen2.5. It outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.";
146
166
  readonly tags: readonly [];
147
167
  readonly properties: {
168
+ readonly context_window: 80000;
148
169
  readonly terms: "https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE";
149
170
  };
150
171
  }, {
@@ -155,6 +176,7 @@ export declare const workersAiCatalog: {
155
176
  readonly tags: readonly [];
156
177
  readonly properties: {
157
178
  readonly beta: true;
179
+ readonly context_window: 4096;
158
180
  readonly info: "https://huggingface.co/TheBloke/DiscoLM_German_7b_v1-AWQ";
159
181
  };
160
182
  }, {
@@ -163,7 +185,9 @@ export declare const workersAiCatalog: {
163
185
  readonly name: "@cf/meta/llama-2-7b-chat-int8";
164
186
  readonly description: "Quantized (int8) generative text model with 7 billion parameters from Meta";
165
187
  readonly tags: readonly [];
166
- readonly properties: {};
188
+ readonly properties: {
189
+ readonly context_window: 8192;
190
+ };
167
191
  }, {
168
192
  readonly id: "9b9c87c6-d4b7-494c-b177-87feab5904db";
169
193
  readonly source: 1;
@@ -171,6 +195,7 @@ export declare const workersAiCatalog: {
171
195
  readonly description: "Llama 3.1 8B quantized to FP8 precision";
172
196
  readonly tags: readonly [];
173
197
  readonly properties: {
198
+ readonly context_window: 32000;
174
199
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
175
200
  };
176
201
  }, {
@@ -181,6 +206,7 @@ export declare const workersAiCatalog: {
181
206
  readonly tags: readonly [];
182
207
  readonly properties: {
183
208
  readonly beta: true;
209
+ readonly context_window: 4096;
184
210
  readonly info: "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-AWQ";
185
211
  };
186
212
  }, {
@@ -191,6 +217,7 @@ export declare const workersAiCatalog: {
191
217
  readonly tags: readonly [];
192
218
  readonly properties: {
193
219
  readonly beta: true;
220
+ readonly context_window: 20000;
194
221
  readonly info: "https://huggingface.co/qwen/qwen1.5-7b-chat-awq";
195
222
  };
196
223
  }, {
@@ -200,6 +227,7 @@ export declare const workersAiCatalog: {
200
227
  readonly description: "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
201
228
  readonly tags: readonly [];
202
229
  readonly properties: {
230
+ readonly context_window: 128000;
203
231
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
204
232
  };
205
233
  }, {
@@ -210,6 +238,7 @@ export declare const workersAiCatalog: {
210
238
  readonly tags: readonly [];
211
239
  readonly properties: {
212
240
  readonly beta: true;
241
+ readonly context_window: 4096;
213
242
  readonly info: "https://huggingface.co/TheBloke/Llama-2-13B-chat-AWQ";
214
243
  };
215
244
  }, {
@@ -220,6 +249,7 @@ export declare const workersAiCatalog: {
220
249
  readonly tags: readonly [];
221
250
  readonly properties: {
222
251
  readonly beta: true;
252
+ readonly context_window: 4096;
223
253
  readonly terms: "https://huggingface.co/TheBloke/deepseek-coder-6.7B-base-AWQ";
224
254
  };
225
255
  }, {
@@ -230,6 +260,7 @@ export declare const workersAiCatalog: {
230
260
  readonly tags: readonly [];
231
261
  readonly properties: {
232
262
  readonly beta: true;
263
+ readonly context_window: 8192;
233
264
  readonly lora: true;
234
265
  };
235
266
  }, {
@@ -239,6 +270,7 @@ export declare const workersAiCatalog: {
239
270
  readonly description: "Llama 3.3 70B quantized to fp8 precision, optimized to be faster.";
240
271
  readonly tags: readonly [];
241
272
  readonly properties: {
273
+ readonly context_window: 24000;
242
274
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_3/LICENSE";
243
275
  };
244
276
  }, {
@@ -249,6 +281,7 @@ export declare const workersAiCatalog: {
249
281
  readonly tags: readonly [];
250
282
  readonly properties: {
251
283
  readonly beta: true;
284
+ readonly context_window: 4096;
252
285
  };
253
286
  }, {
254
287
  readonly id: "60474554-f03b-4ff4-8ecc-c1b7c71d7b29";
@@ -258,6 +291,7 @@ export declare const workersAiCatalog: {
258
291
  readonly tags: readonly [];
259
292
  readonly properties: {
260
293
  readonly beta: true;
294
+ readonly context_window: 4096;
261
295
  readonly terms: "https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-AWQ";
262
296
  };
263
297
  }, {
@@ -268,6 +302,7 @@ export declare const workersAiCatalog: {
268
302
  readonly tags: readonly [];
269
303
  readonly properties: {
270
304
  readonly beta: true;
305
+ readonly context_window: 4096;
271
306
  readonly info: "https://huggingface.co/deepseek-ai/deepseek-math-7b-instruct";
272
307
  readonly terms: "https://github.com/deepseek-ai/DeepSeek-Math/blob/main/LICENSE-MODEL";
273
308
  };
@@ -279,6 +314,7 @@ export declare const workersAiCatalog: {
279
314
  readonly tags: readonly [];
280
315
  readonly properties: {
281
316
  readonly beta: true;
317
+ readonly context_window: 4096;
282
318
  readonly info: "https://huggingface.co/tiiuae/falcon-7b-instruct";
283
319
  };
284
320
  }, {
@@ -289,6 +325,7 @@ export declare const workersAiCatalog: {
289
325
  readonly tags: readonly [];
290
326
  readonly properties: {
291
327
  readonly beta: true;
328
+ readonly context_window: 24000;
292
329
  readonly function_calling: true;
293
330
  readonly info: "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B";
294
331
  };
@@ -299,6 +336,7 @@ export declare const workersAiCatalog: {
299
336
  readonly description: "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.";
300
337
  readonly tags: readonly [];
301
338
  readonly properties: {
339
+ readonly context_window: 7968;
302
340
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
303
341
  };
304
342
  }, {
@@ -308,6 +346,7 @@ export declare const workersAiCatalog: {
308
346
  readonly description: "Quantized (int4) generative text model with 8 billion parameters from Meta.\n";
309
347
  readonly tags: readonly [];
310
348
  readonly properties: {
349
+ readonly context_window: 8192;
311
350
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
312
351
  };
313
352
  }, {
@@ -318,6 +357,7 @@ export declare const workersAiCatalog: {
318
357
  readonly tags: readonly [];
319
358
  readonly properties: {
320
359
  readonly beta: true;
360
+ readonly context_window: 4096;
321
361
  readonly info: "https://huggingface.co/TheBloke/zephyr-7B-beta-AWQ";
322
362
  };
323
363
  }, {
@@ -328,6 +368,7 @@ export declare const workersAiCatalog: {
328
368
  readonly tags: readonly [];
329
369
  readonly properties: {
330
370
  readonly beta: true;
371
+ readonly context_window: 3500;
331
372
  readonly lora: true;
332
373
  };
333
374
  }, {
@@ -338,6 +379,7 @@ export declare const workersAiCatalog: {
338
379
  readonly tags: readonly [];
339
380
  readonly properties: {
340
381
  readonly beta: true;
382
+ readonly context_window: 32000;
341
383
  readonly info: "https://huggingface.co/qwen/qwen1.5-1.8b-chat";
342
384
  };
343
385
  }, {
@@ -347,6 +389,7 @@ export declare const workersAiCatalog: {
347
389
  readonly description: "Quantized (int4) generative text model with 8 billion parameters from Meta.";
348
390
  readonly tags: readonly [];
349
391
  readonly properties: {
392
+ readonly context_window: 8192;
350
393
  readonly info: "https://llama.meta.com";
351
394
  readonly terms: "https://llama.meta.com/llama3/license/#";
352
395
  };
@@ -367,6 +410,7 @@ export declare const workersAiCatalog: {
367
410
  readonly tags: readonly [];
368
411
  readonly properties: {
369
412
  readonly beta: true;
413
+ readonly context_window: 10000;
370
414
  readonly info: "https://huggingface.co/defog/sqlcoder-7b-2";
371
415
  readonly terms: "https://creativecommons.org/licenses/by-sa/4.0/deed.en";
372
416
  };
@@ -378,6 +422,7 @@ export declare const workersAiCatalog: {
378
422
  readonly tags: readonly [];
379
423
  readonly properties: {
380
424
  readonly beta: true;
425
+ readonly context_window: 2048;
381
426
  readonly info: "https://huggingface.co/microsoft/phi-2";
382
427
  };
383
428
  }, {
@@ -386,7 +431,9 @@ export declare const workersAiCatalog: {
386
431
  readonly name: "@hf/meta-llama/meta-llama-3-8b-instruct";
387
432
  readonly description: "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.\t";
388
433
  readonly tags: readonly [];
389
- readonly properties: {};
434
+ readonly properties: {
435
+ readonly context_window: 8192;
436
+ };
390
437
  }, {
391
438
  readonly id: "0f002249-7d86-4698-aabf-8529ed86cefb";
392
439
  readonly source: 2;
@@ -395,11 +442,9 @@ export declare const workersAiCatalog: {
395
442
  readonly tags: readonly [];
396
443
  readonly properties: {
397
444
  readonly beta: true;
445
+ readonly context_window: 8192;
398
446
  readonly info: "https://ai.google.dev/gemma/docs";
399
447
  readonly lora: true;
400
- readonly max_batch_prefill_tokens: 2048;
401
- readonly max_input_length: 1512;
402
- readonly max_total_tokens: 2048;
403
448
  readonly terms: "https://ai.google.dev/gemma/terms";
404
449
  };
405
450
  }, {
@@ -410,6 +455,7 @@ export declare const workersAiCatalog: {
410
455
  readonly tags: readonly [];
411
456
  readonly properties: {
412
457
  readonly beta: true;
458
+ readonly context_window: 7500;
413
459
  readonly info: "https://huggingface.co/qwen/qwen1.5-14b-chat-awq";
414
460
  };
415
461
  }, {
@@ -420,6 +466,7 @@ export declare const workersAiCatalog: {
420
466
  readonly tags: readonly [];
421
467
  readonly properties: {
422
468
  readonly beta: true;
469
+ readonly context_window: 8192;
423
470
  readonly info: "https://huggingface.co/openchat/openchat-3.5-0106";
424
471
  };
425
472
  }];
@@ -12,6 +12,7 @@ export const workersAiCatalog = {
12
12
  tags: [],
13
13
  properties: {
14
14
  beta: true,
15
+ context_window: 32000,
15
16
  info: 'https://huggingface.co/qwen/qwen1.5-0.5b-chat',
16
17
  },
17
18
  },
@@ -23,6 +24,7 @@ export const workersAiCatalog = {
23
24
  tags: [],
24
25
  properties: {
25
26
  beta: true,
27
+ context_window: 8192,
26
28
  lora: true,
27
29
  },
28
30
  },
@@ -34,6 +36,7 @@ export const workersAiCatalog = {
34
36
  tags: [],
35
37
  properties: {
36
38
  beta: true,
39
+ context_window: 4096,
37
40
  info: 'https://huggingface.co/Nexusflow/Starling-LM-7B-beta',
38
41
  max_batch_prefill_tokens: 8192,
39
42
  max_input_length: 3072,
@@ -47,6 +50,7 @@ export const workersAiCatalog = {
47
50
  description: 'Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.',
48
51
  tags: [],
49
52
  properties: {
53
+ context_window: 7968,
50
54
  info: 'https://llama.meta.com',
51
55
  terms: 'https://llama.meta.com/llama3/license/#',
52
56
  },
@@ -58,6 +62,7 @@ export const workersAiCatalog = {
58
62
  description: 'The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.',
59
63
  tags: [],
60
64
  properties: {
65
+ context_window: 128000,
61
66
  terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE',
62
67
  },
63
68
  },
@@ -69,6 +74,7 @@ export const workersAiCatalog = {
69
74
  tags: [],
70
75
  properties: {
71
76
  beta: true,
77
+ context_window: 4096,
72
78
  },
73
79
  },
74
80
  {
@@ -79,8 +85,17 @@ export const workersAiCatalog = {
79
85
  tags: [],
80
86
  properties: {
81
87
  beta: true,
88
+ context_window: 4096,
82
89
  },
83
90
  },
91
+ {
92
+ id: 'cc80437b-9a8d-4f1a-9c77-9aaf0d226922',
93
+ source: 1,
94
+ name: '@cf/meta/llama-guard-3-8b',
95
+ description: 'Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.',
96
+ tags: [],
97
+ properties: {},
98
+ },
84
99
  {
85
100
  id: 'ca54bcd6-0d98-4739-9b3b-5c8b4402193d',
86
101
  source: 1,
@@ -89,6 +104,7 @@ export const workersAiCatalog = {
89
104
  tags: [],
90
105
  properties: {
91
106
  beta: true,
107
+ context_window: 4096,
92
108
  info: 'https://ai.meta.com/llama/',
93
109
  terms: 'https://ai.meta.com/resources/models-and-libraries/llama-downloads/',
94
110
  },
@@ -101,6 +117,7 @@ export const workersAiCatalog = {
101
117
  tags: [],
102
118
  properties: {
103
119
  beta: true,
120
+ context_window: 2824,
104
121
  info: 'https://mistral.ai/news/announcing-mistral-7b/',
105
122
  lora: true,
106
123
  },
@@ -113,6 +130,7 @@ export const workersAiCatalog = {
113
130
  tags: [],
114
131
  properties: {
115
132
  beta: true,
133
+ context_window: 15000,
116
134
  lora: true,
117
135
  },
118
136
  },
@@ -124,6 +142,7 @@ export const workersAiCatalog = {
124
142
  tags: [],
125
143
  properties: {
126
144
  beta: true,
145
+ context_window: 2048,
127
146
  info: 'https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0',
128
147
  },
129
148
  },
@@ -135,6 +154,7 @@ export const workersAiCatalog = {
135
154
  tags: [],
136
155
  properties: {
137
156
  beta: true,
157
+ context_window: 3072,
138
158
  info: 'https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2',
139
159
  lora: true,
140
160
  max_batch_prefill_tokens: 8192,
@@ -150,6 +170,7 @@ export const workersAiCatalog = {
150
170
  tags: [],
151
171
  properties: {
152
172
  beta: true,
173
+ context_window: 15000,
153
174
  },
154
175
  },
155
176
  {
@@ -159,6 +180,7 @@ export const workersAiCatalog = {
159
180
  description: 'DeepSeek-R1-Distill-Qwen-32B is a model distilled from DeepSeek-R1 based on Qwen2.5. It outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.',
160
181
  tags: [],
161
182
  properties: {
183
+ context_window: 80000,
162
184
  terms: 'https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE',
163
185
  },
164
186
  },
@@ -170,6 +192,7 @@ export const workersAiCatalog = {
170
192
  tags: [],
171
193
  properties: {
172
194
  beta: true,
195
+ context_window: 4096,
173
196
  info: 'https://huggingface.co/TheBloke/DiscoLM_German_7b_v1-AWQ',
174
197
  },
175
198
  },
@@ -179,7 +202,9 @@ export const workersAiCatalog = {
179
202
  name: '@cf/meta/llama-2-7b-chat-int8',
180
203
  description: 'Quantized (int8) generative text model with 7 billion parameters from Meta',
181
204
  tags: [],
182
- properties: {},
205
+ properties: {
206
+ context_window: 8192,
207
+ },
183
208
  },
184
209
  {
185
210
  id: '9b9c87c6-d4b7-494c-b177-87feab5904db',
@@ -188,6 +213,7 @@ export const workersAiCatalog = {
188
213
  description: 'Llama 3.1 8B quantized to FP8 precision',
189
214
  tags: [],
190
215
  properties: {
216
+ context_window: 32000,
191
217
  terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE',
192
218
  },
193
219
  },
@@ -199,6 +225,7 @@ export const workersAiCatalog = {
199
225
  tags: [],
200
226
  properties: {
201
227
  beta: true,
228
+ context_window: 4096,
202
229
  info: 'https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-AWQ',
203
230
  },
204
231
  },
@@ -210,6 +237,7 @@ export const workersAiCatalog = {
210
237
  tags: [],
211
238
  properties: {
212
239
  beta: true,
240
+ context_window: 20000,
213
241
  info: 'https://huggingface.co/qwen/qwen1.5-7b-chat-awq',
214
242
  },
215
243
  },
@@ -220,6 +248,7 @@ export const workersAiCatalog = {
220
248
  description: 'The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.',
221
249
  tags: [],
222
250
  properties: {
251
+ context_window: 128000,
223
252
  terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE',
224
253
  },
225
254
  },
@@ -231,6 +260,7 @@ export const workersAiCatalog = {
231
260
  tags: [],
232
261
  properties: {
233
262
  beta: true,
263
+ context_window: 4096,
234
264
  info: 'https://huggingface.co/TheBloke/Llama-2-13B-chat-AWQ',
235
265
  },
236
266
  },
@@ -242,6 +272,7 @@ export const workersAiCatalog = {
242
272
  tags: [],
243
273
  properties: {
244
274
  beta: true,
275
+ context_window: 4096,
245
276
  terms: 'https://huggingface.co/TheBloke/deepseek-coder-6.7B-base-AWQ',
246
277
  },
247
278
  },
@@ -253,6 +284,7 @@ export const workersAiCatalog = {
253
284
  tags: [],
254
285
  properties: {
255
286
  beta: true,
287
+ context_window: 8192,
256
288
  lora: true,
257
289
  },
258
290
  },
@@ -263,6 +295,7 @@ export const workersAiCatalog = {
263
295
  description: 'Llama 3.3 70B quantized to fp8 precision, optimized to be faster.',
264
296
  tags: [],
265
297
  properties: {
298
+ context_window: 24000,
266
299
  terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_3/LICENSE',
267
300
  },
268
301
  },
@@ -274,6 +307,7 @@ export const workersAiCatalog = {
274
307
  tags: [],
275
308
  properties: {
276
309
  beta: true,
310
+ context_window: 4096,
277
311
  },
278
312
  },
279
313
  {
@@ -284,6 +318,7 @@ export const workersAiCatalog = {
284
318
  tags: [],
285
319
  properties: {
286
320
  beta: true,
321
+ context_window: 4096,
287
322
  terms: 'https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-AWQ',
288
323
  },
289
324
  },
@@ -295,6 +330,7 @@ export const workersAiCatalog = {
295
330
  tags: [],
296
331
  properties: {
297
332
  beta: true,
333
+ context_window: 4096,
298
334
  info: 'https://huggingface.co/deepseek-ai/deepseek-math-7b-instruct',
299
335
  terms: 'https://github.com/deepseek-ai/DeepSeek-Math/blob/main/LICENSE-MODEL',
300
336
  },
@@ -307,6 +343,7 @@ export const workersAiCatalog = {
307
343
  tags: [],
308
344
  properties: {
309
345
  beta: true,
346
+ context_window: 4096,
310
347
  info: 'https://huggingface.co/tiiuae/falcon-7b-instruct',
311
348
  },
312
349
  },
@@ -318,6 +355,7 @@ export const workersAiCatalog = {
318
355
  tags: [],
319
356
  properties: {
320
357
  beta: true,
358
+ context_window: 24000,
321
359
  function_calling: true,
322
360
  info: 'https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B',
323
361
  },
@@ -329,6 +367,7 @@ export const workersAiCatalog = {
329
367
  description: 'The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.',
330
368
  tags: [],
331
369
  properties: {
370
+ context_window: 7968,
332
371
  terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE',
333
372
  },
334
373
  },
@@ -339,6 +378,7 @@ export const workersAiCatalog = {
339
378
  description: 'Quantized (int4) generative text model with 8 billion parameters from Meta.\n',
340
379
  tags: [],
341
380
  properties: {
381
+ context_window: 8192,
342
382
  terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE',
343
383
  },
344
384
  },
@@ -350,6 +390,7 @@ export const workersAiCatalog = {
350
390
  tags: [],
351
391
  properties: {
352
392
  beta: true,
393
+ context_window: 4096,
353
394
  info: 'https://huggingface.co/TheBloke/zephyr-7B-beta-AWQ',
354
395
  },
355
396
  },
@@ -361,6 +402,7 @@ export const workersAiCatalog = {
361
402
  tags: [],
362
403
  properties: {
363
404
  beta: true,
405
+ context_window: 3500,
364
406
  lora: true,
365
407
  },
366
408
  },
@@ -372,6 +414,7 @@ export const workersAiCatalog = {
372
414
  tags: [],
373
415
  properties: {
374
416
  beta: true,
417
+ context_window: 32000,
375
418
  info: 'https://huggingface.co/qwen/qwen1.5-1.8b-chat',
376
419
  },
377
420
  },
@@ -382,6 +425,7 @@ export const workersAiCatalog = {
382
425
  description: 'Quantized (int4) generative text model with 8 billion parameters from Meta.',
383
426
  tags: [],
384
427
  properties: {
428
+ context_window: 8192,
385
429
  info: 'https://llama.meta.com',
386
430
  terms: 'https://llama.meta.com/llama3/license/#',
387
431
  },
@@ -404,6 +448,7 @@ export const workersAiCatalog = {
404
448
  tags: [],
405
449
  properties: {
406
450
  beta: true,
451
+ context_window: 10000,
407
452
  info: 'https://huggingface.co/defog/sqlcoder-7b-2',
408
453
  terms: 'https://creativecommons.org/licenses/by-sa/4.0/deed.en',
409
454
  },
@@ -416,6 +461,7 @@ export const workersAiCatalog = {
416
461
  tags: [],
417
462
  properties: {
418
463
  beta: true,
464
+ context_window: 2048,
419
465
  info: 'https://huggingface.co/microsoft/phi-2',
420
466
  },
421
467
  },
@@ -425,7 +471,9 @@ export const workersAiCatalog = {
425
471
  name: '@hf/meta-llama/meta-llama-3-8b-instruct',
426
472
  description: 'Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.\t',
427
473
  tags: [],
428
- properties: {},
474
+ properties: {
475
+ context_window: 8192,
476
+ },
429
477
  },
430
478
  {
431
479
  id: '0f002249-7d86-4698-aabf-8529ed86cefb',
@@ -435,11 +483,9 @@ export const workersAiCatalog = {
435
483
  tags: [],
436
484
  properties: {
437
485
  beta: true,
486
+ context_window: 8192,
438
487
  info: 'https://ai.google.dev/gemma/docs',
439
488
  lora: true,
440
- max_batch_prefill_tokens: 2048,
441
- max_input_length: 1512,
442
- max_total_tokens: 2048,
443
489
  terms: 'https://ai.google.dev/gemma/terms',
444
490
  },
445
491
  },
@@ -451,6 +497,7 @@ export const workersAiCatalog = {
451
497
  tags: [],
452
498
  properties: {
453
499
  beta: true,
500
+ context_window: 7500,
454
501
  info: 'https://huggingface.co/qwen/qwen1.5-14b-chat-awq',
455
502
  },
456
503
  },
@@ -462,6 +509,7 @@ export const workersAiCatalog = {
462
509
  tags: [],
463
510
  properties: {
464
511
  beta: true,
512
+ context_window: 8192,
465
513
  info: 'https://huggingface.co/openchat/openchat-3.5-0106',
466
514
  },
467
515
  },
@@ -50,6 +50,7 @@ export declare const possibilities_mc_generic: readonly [...{
50
50
  readonly tags: readonly [];
51
51
  readonly properties: {
52
52
  readonly beta: true;
53
+ readonly context_window: 24000;
53
54
  readonly function_calling: true;
54
55
  readonly info: "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B";
55
56
  };
@@ -78,6 +79,7 @@ export declare const possibilities_mc_extraction: readonly [...{
78
79
  readonly tags: readonly [];
79
80
  readonly properties: {
80
81
  readonly beta: true;
82
+ readonly context_window: 24000;
81
83
  readonly function_calling: true;
82
84
  readonly info: "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B";
83
85
  };
@@ -93,6 +95,7 @@ export declare const possibilities_mc_tagging: readonly [...{
93
95
  readonly tags: readonly [];
94
96
  readonly properties: {
95
97
  readonly beta: true;
98
+ readonly context_window: 24000;
96
99
  readonly function_calling: true;
97
100
  readonly info: "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B";
98
101
  };
@@ -108,6 +111,7 @@ export declare const possibilities_mc_sentiment: readonly [...{
108
111
  readonly tags: readonly [];
109
112
  readonly properties: {
110
113
  readonly beta: true;
114
+ readonly context_window: 24000;
111
115
  readonly function_calling: true;
112
116
  readonly info: "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B";
113
117
  };
@@ -123,6 +127,7 @@ export declare const possibilities_mc_safety: readonly [...{
123
127
  readonly tags: readonly [];
124
128
  readonly properties: {
125
129
  readonly beta: true;
130
+ readonly context_window: 32000;
126
131
  readonly info: "https://huggingface.co/qwen/qwen1.5-0.5b-chat";
127
132
  };
128
133
  }, {
@@ -133,6 +138,7 @@ export declare const possibilities_mc_safety: readonly [...{
133
138
  readonly tags: readonly [];
134
139
  readonly properties: {
135
140
  readonly beta: true;
141
+ readonly context_window: 8192;
136
142
  readonly lora: true;
137
143
  };
138
144
  }, {
@@ -143,6 +149,7 @@ export declare const possibilities_mc_safety: readonly [...{
143
149
  readonly tags: readonly [];
144
150
  readonly properties: {
145
151
  readonly beta: true;
152
+ readonly context_window: 4096;
146
153
  readonly info: "https://huggingface.co/Nexusflow/Starling-LM-7B-beta";
147
154
  readonly max_batch_prefill_tokens: 8192;
148
155
  readonly max_input_length: 3072;
@@ -155,6 +162,7 @@ export declare const possibilities_mc_safety: readonly [...{
155
162
  readonly description: "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.";
156
163
  readonly tags: readonly [];
157
164
  readonly properties: {
165
+ readonly context_window: 7968;
158
166
  readonly info: "https://llama.meta.com";
159
167
  readonly terms: "https://llama.meta.com/llama3/license/#";
160
168
  };
@@ -165,6 +173,7 @@ export declare const possibilities_mc_safety: readonly [...{
165
173
  readonly description: "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
166
174
  readonly tags: readonly [];
167
175
  readonly properties: {
176
+ readonly context_window: 128000;
168
177
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
169
178
  };
170
179
  }, {
@@ -175,6 +184,7 @@ export declare const possibilities_mc_safety: readonly [...{
175
184
  readonly tags: readonly [];
176
185
  readonly properties: {
177
186
  readonly beta: true;
187
+ readonly context_window: 4096;
178
188
  };
179
189
  }, {
180
190
  readonly id: "d2ba5c6b-bbb7-49d6-b466-900654870cd6";
@@ -184,7 +194,15 @@ export declare const possibilities_mc_safety: readonly [...{
184
194
  readonly tags: readonly [];
185
195
  readonly properties: {
186
196
  readonly beta: true;
197
+ readonly context_window: 4096;
187
198
  };
199
+ }, {
200
+ readonly id: "cc80437b-9a8d-4f1a-9c77-9aaf0d226922";
201
+ readonly source: 1;
202
+ readonly name: "@cf/meta/llama-guard-3-8b";
203
+ readonly description: "Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.";
204
+ readonly tags: readonly [];
205
+ readonly properties: {};
188
206
  }, {
189
207
  readonly id: "ca54bcd6-0d98-4739-9b3b-5c8b4402193d";
190
208
  readonly source: 1;
@@ -193,6 +211,7 @@ export declare const possibilities_mc_safety: readonly [...{
193
211
  readonly tags: readonly [];
194
212
  readonly properties: {
195
213
  readonly beta: true;
214
+ readonly context_window: 4096;
196
215
  readonly info: "https://ai.meta.com/llama/";
197
216
  readonly terms: "https://ai.meta.com/resources/models-and-libraries/llama-downloads/";
198
217
  };
@@ -204,6 +223,7 @@ export declare const possibilities_mc_safety: readonly [...{
204
223
  readonly tags: readonly [];
205
224
  readonly properties: {
206
225
  readonly beta: true;
226
+ readonly context_window: 2824;
207
227
  readonly info: "https://mistral.ai/news/announcing-mistral-7b/";
208
228
  readonly lora: true;
209
229
  };
@@ -215,6 +235,7 @@ export declare const possibilities_mc_safety: readonly [...{
215
235
  readonly tags: readonly [];
216
236
  readonly properties: {
217
237
  readonly beta: true;
238
+ readonly context_window: 15000;
218
239
  readonly lora: true;
219
240
  };
220
241
  }, {
@@ -225,6 +246,7 @@ export declare const possibilities_mc_safety: readonly [...{
225
246
  readonly tags: readonly [];
226
247
  readonly properties: {
227
248
  readonly beta: true;
249
+ readonly context_window: 2048;
228
250
  readonly info: "https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0";
229
251
  };
230
252
  }, {
@@ -235,6 +257,7 @@ export declare const possibilities_mc_safety: readonly [...{
235
257
  readonly tags: readonly [];
236
258
  readonly properties: {
237
259
  readonly beta: true;
260
+ readonly context_window: 3072;
238
261
  readonly info: "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2";
239
262
  readonly lora: true;
240
263
  readonly max_batch_prefill_tokens: 8192;
@@ -249,6 +272,7 @@ export declare const possibilities_mc_safety: readonly [...{
249
272
  readonly tags: readonly [];
250
273
  readonly properties: {
251
274
  readonly beta: true;
275
+ readonly context_window: 15000;
252
276
  };
253
277
  }, {
254
278
  readonly id: "ad01ab83-baf8-4e7b-8fed-a0a219d4eb45";
@@ -257,6 +281,7 @@ export declare const possibilities_mc_safety: readonly [...{
257
281
  readonly description: "DeepSeek-R1-Distill-Qwen-32B is a model distilled from DeepSeek-R1 based on Qwen2.5. It outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.";
258
282
  readonly tags: readonly [];
259
283
  readonly properties: {
284
+ readonly context_window: 80000;
260
285
  readonly terms: "https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE";
261
286
  };
262
287
  }, {
@@ -267,6 +292,7 @@ export declare const possibilities_mc_safety: readonly [...{
267
292
  readonly tags: readonly [];
268
293
  readonly properties: {
269
294
  readonly beta: true;
295
+ readonly context_window: 4096;
270
296
  readonly info: "https://huggingface.co/TheBloke/DiscoLM_German_7b_v1-AWQ";
271
297
  };
272
298
  }, {
@@ -275,7 +301,9 @@ export declare const possibilities_mc_safety: readonly [...{
275
301
  readonly name: "@cf/meta/llama-2-7b-chat-int8";
276
302
  readonly description: "Quantized (int8) generative text model with 7 billion parameters from Meta";
277
303
  readonly tags: readonly [];
278
- readonly properties: {};
304
+ readonly properties: {
305
+ readonly context_window: 8192;
306
+ };
279
307
  }, {
280
308
  readonly id: "9b9c87c6-d4b7-494c-b177-87feab5904db";
281
309
  readonly source: 1;
@@ -283,6 +311,7 @@ export declare const possibilities_mc_safety: readonly [...{
283
311
  readonly description: "Llama 3.1 8B quantized to FP8 precision";
284
312
  readonly tags: readonly [];
285
313
  readonly properties: {
314
+ readonly context_window: 32000;
286
315
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
287
316
  };
288
317
  }, {
@@ -293,6 +322,7 @@ export declare const possibilities_mc_safety: readonly [...{
293
322
  readonly tags: readonly [];
294
323
  readonly properties: {
295
324
  readonly beta: true;
325
+ readonly context_window: 4096;
296
326
  readonly info: "https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-AWQ";
297
327
  };
298
328
  }, {
@@ -303,6 +333,7 @@ export declare const possibilities_mc_safety: readonly [...{
303
333
  readonly tags: readonly [];
304
334
  readonly properties: {
305
335
  readonly beta: true;
336
+ readonly context_window: 20000;
306
337
  readonly info: "https://huggingface.co/qwen/qwen1.5-7b-chat-awq";
307
338
  };
308
339
  }, {
@@ -312,6 +343,7 @@ export declare const possibilities_mc_safety: readonly [...{
312
343
  readonly description: "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
313
344
  readonly tags: readonly [];
314
345
  readonly properties: {
346
+ readonly context_window: 128000;
315
347
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
316
348
  };
317
349
  }, {
@@ -322,6 +354,7 @@ export declare const possibilities_mc_safety: readonly [...{
322
354
  readonly tags: readonly [];
323
355
  readonly properties: {
324
356
  readonly beta: true;
357
+ readonly context_window: 4096;
325
358
  readonly info: "https://huggingface.co/TheBloke/Llama-2-13B-chat-AWQ";
326
359
  };
327
360
  }, {
@@ -332,6 +365,7 @@ export declare const possibilities_mc_safety: readonly [...{
332
365
  readonly tags: readonly [];
333
366
  readonly properties: {
334
367
  readonly beta: true;
368
+ readonly context_window: 4096;
335
369
  readonly terms: "https://huggingface.co/TheBloke/deepseek-coder-6.7B-base-AWQ";
336
370
  };
337
371
  }, {
@@ -342,6 +376,7 @@ export declare const possibilities_mc_safety: readonly [...{
342
376
  readonly tags: readonly [];
343
377
  readonly properties: {
344
378
  readonly beta: true;
379
+ readonly context_window: 8192;
345
380
  readonly lora: true;
346
381
  };
347
382
  }, {
@@ -351,6 +386,7 @@ export declare const possibilities_mc_safety: readonly [...{
351
386
  readonly description: "Llama 3.3 70B quantized to fp8 precision, optimized to be faster.";
352
387
  readonly tags: readonly [];
353
388
  readonly properties: {
389
+ readonly context_window: 24000;
354
390
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_3/LICENSE";
355
391
  };
356
392
  }, {
@@ -361,6 +397,7 @@ export declare const possibilities_mc_safety: readonly [...{
361
397
  readonly tags: readonly [];
362
398
  readonly properties: {
363
399
  readonly beta: true;
400
+ readonly context_window: 4096;
364
401
  };
365
402
  }, {
366
403
  readonly id: "60474554-f03b-4ff4-8ecc-c1b7c71d7b29";
@@ -370,6 +407,7 @@ export declare const possibilities_mc_safety: readonly [...{
370
407
  readonly tags: readonly [];
371
408
  readonly properties: {
372
409
  readonly beta: true;
410
+ readonly context_window: 4096;
373
411
  readonly terms: "https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-AWQ";
374
412
  };
375
413
  }, {
@@ -380,6 +418,7 @@ export declare const possibilities_mc_safety: readonly [...{
380
418
  readonly tags: readonly [];
381
419
  readonly properties: {
382
420
  readonly beta: true;
421
+ readonly context_window: 4096;
383
422
  readonly info: "https://huggingface.co/deepseek-ai/deepseek-math-7b-instruct";
384
423
  readonly terms: "https://github.com/deepseek-ai/DeepSeek-Math/blob/main/LICENSE-MODEL";
385
424
  };
@@ -391,6 +430,7 @@ export declare const possibilities_mc_safety: readonly [...{
391
430
  readonly tags: readonly [];
392
431
  readonly properties: {
393
432
  readonly beta: true;
433
+ readonly context_window: 4096;
394
434
  readonly info: "https://huggingface.co/tiiuae/falcon-7b-instruct";
395
435
  };
396
436
  }, {
@@ -401,6 +441,7 @@ export declare const possibilities_mc_safety: readonly [...{
401
441
  readonly tags: readonly [];
402
442
  readonly properties: {
403
443
  readonly beta: true;
444
+ readonly context_window: 24000;
404
445
  readonly function_calling: true;
405
446
  readonly info: "https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B";
406
447
  };
@@ -411,6 +452,7 @@ export declare const possibilities_mc_safety: readonly [...{
411
452
  readonly description: "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.";
412
453
  readonly tags: readonly [];
413
454
  readonly properties: {
455
+ readonly context_window: 7968;
414
456
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
415
457
  };
416
458
  }, {
@@ -420,6 +462,7 @@ export declare const possibilities_mc_safety: readonly [...{
420
462
  readonly description: "Quantized (int4) generative text model with 8 billion parameters from Meta.\n";
421
463
  readonly tags: readonly [];
422
464
  readonly properties: {
465
+ readonly context_window: 8192;
423
466
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
424
467
  };
425
468
  }, {
@@ -430,6 +473,7 @@ export declare const possibilities_mc_safety: readonly [...{
430
473
  readonly tags: readonly [];
431
474
  readonly properties: {
432
475
  readonly beta: true;
476
+ readonly context_window: 4096;
433
477
  readonly info: "https://huggingface.co/TheBloke/zephyr-7B-beta-AWQ";
434
478
  };
435
479
  }, {
@@ -440,6 +484,7 @@ export declare const possibilities_mc_safety: readonly [...{
440
484
  readonly tags: readonly [];
441
485
  readonly properties: {
442
486
  readonly beta: true;
487
+ readonly context_window: 3500;
443
488
  readonly lora: true;
444
489
  };
445
490
  }, {
@@ -450,6 +495,7 @@ export declare const possibilities_mc_safety: readonly [...{
450
495
  readonly tags: readonly [];
451
496
  readonly properties: {
452
497
  readonly beta: true;
498
+ readonly context_window: 32000;
453
499
  readonly info: "https://huggingface.co/qwen/qwen1.5-1.8b-chat";
454
500
  };
455
501
  }, {
@@ -459,6 +505,7 @@ export declare const possibilities_mc_safety: readonly [...{
459
505
  readonly description: "Quantized (int4) generative text model with 8 billion parameters from Meta.";
460
506
  readonly tags: readonly [];
461
507
  readonly properties: {
508
+ readonly context_window: 8192;
462
509
  readonly info: "https://llama.meta.com";
463
510
  readonly terms: "https://llama.meta.com/llama3/license/#";
464
511
  };
@@ -479,6 +526,7 @@ export declare const possibilities_mc_safety: readonly [...{
479
526
  readonly tags: readonly [];
480
527
  readonly properties: {
481
528
  readonly beta: true;
529
+ readonly context_window: 10000;
482
530
  readonly info: "https://huggingface.co/defog/sqlcoder-7b-2";
483
531
  readonly terms: "https://creativecommons.org/licenses/by-sa/4.0/deed.en";
484
532
  };
@@ -490,6 +538,7 @@ export declare const possibilities_mc_safety: readonly [...{
490
538
  readonly tags: readonly [];
491
539
  readonly properties: {
492
540
  readonly beta: true;
541
+ readonly context_window: 2048;
493
542
  readonly info: "https://huggingface.co/microsoft/phi-2";
494
543
  };
495
544
  }, {
@@ -498,7 +547,9 @@ export declare const possibilities_mc_safety: readonly [...{
498
547
  readonly name: "@hf/meta-llama/meta-llama-3-8b-instruct";
499
548
  readonly description: "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.\t";
500
549
  readonly tags: readonly [];
501
- readonly properties: {};
550
+ readonly properties: {
551
+ readonly context_window: 8192;
552
+ };
502
553
  }, {
503
554
  readonly id: "0f002249-7d86-4698-aabf-8529ed86cefb";
504
555
  readonly source: 2;
@@ -507,11 +558,9 @@ export declare const possibilities_mc_safety: readonly [...{
507
558
  readonly tags: readonly [];
508
559
  readonly properties: {
509
560
  readonly beta: true;
561
+ readonly context_window: 8192;
510
562
  readonly info: "https://ai.google.dev/gemma/docs";
511
563
  readonly lora: true;
512
- readonly max_batch_prefill_tokens: 2048;
513
- readonly max_input_length: 1512;
514
- readonly max_total_tokens: 2048;
515
564
  readonly terms: "https://ai.google.dev/gemma/terms";
516
565
  };
517
566
  }, {
@@ -522,6 +571,7 @@ export declare const possibilities_mc_safety: readonly [...{
522
571
  readonly tags: readonly [];
523
572
  readonly properties: {
524
573
  readonly beta: true;
574
+ readonly context_window: 7500;
525
575
  readonly info: "https://huggingface.co/qwen/qwen1.5-14b-chat-awq";
526
576
  };
527
577
  }, {
@@ -532,6 +582,7 @@ export declare const possibilities_mc_safety: readonly [...{
532
582
  readonly tags: readonly [];
533
583
  readonly properties: {
534
584
  readonly beta: true;
585
+ readonly context_window: 8192;
535
586
  readonly info: "https://huggingface.co/openchat/openchat-3.5-0106";
536
587
  };
537
588
  }];
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@chainfuse/types",
3
- "version": "1.6.7",
3
+ "version": "1.6.9",
4
4
  "description": "",
5
5
  "author": "ChainFuse",
6
6
  "homepage": "https://github.com/ChainFuse/packages/tree/main/packages/types#readme",
@@ -86,8 +86,8 @@
86
86
  },
87
87
  "prettier": "@demosjarco/prettier-config",
88
88
  "devDependencies": {
89
- "@cloudflare/workers-types": "^4.20250214.0",
89
+ "@cloudflare/workers-types": "^4.20250224.0",
90
90
  "@types/json-schema": "^7.0.15"
91
91
  },
92
- "gitHead": "4ea3e6aed331ba9832092618ef88fa33cf38ca64"
92
+ "gitHead": "9d842a11f77e62dab485f3965c724827b44f688f"
93
93
  }