@chainfuse/types 2.2.0 → 2.2.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -8,6 +8,7 @@ export declare const workersAiCatalog: {
8
8
  readonly source: 1;
9
9
  readonly name: "@cf/qwen/qwen1.5-0.5b-chat";
10
10
  readonly description: "Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud.";
11
+ readonly created_at: "2024-02-27 18:23:37.344";
11
12
  readonly tags: readonly [];
12
13
  readonly properties: {
13
14
  readonly beta: true;
@@ -19,6 +20,7 @@ export declare const workersAiCatalog: {
19
20
  readonly source: 1;
20
21
  readonly name: "@cf/google/gemma-2b-it-lora";
21
22
  readonly description: "This is a Gemma-2B base model that Cloudflare dedicates for inference with LoRA adapters. Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.";
23
+ readonly created_at: "2024-04-02 00:19:34.669";
22
24
  readonly tags: readonly [];
23
25
  readonly properties: {
24
26
  readonly beta: true;
@@ -30,6 +32,7 @@ export declare const workersAiCatalog: {
30
32
  readonly source: 2;
31
33
  readonly name: "@hf/nexusflow/starling-lm-7b-beta";
32
34
  readonly description: "We introduce Starling-LM-7B-beta, an open large language model (LLM) trained by Reinforcement Learning from AI Feedback (RLAIF). Starling-LM-7B-beta is trained from Openchat-3.5-0106 with our new reward model Nexusflow/Starling-RM-34B and policy optimization method Fine-Tuning Language Models from Human Preferences (PPO).";
35
+ readonly created_at: "2024-04-01 23:49:31.797";
33
36
  readonly tags: readonly [];
34
37
  readonly properties: {
35
38
  readonly beta: true;
@@ -44,8 +47,16 @@ export declare const workersAiCatalog: {
44
47
  readonly source: 1;
45
48
  readonly name: "@cf/meta/llama-3-8b-instruct";
46
49
  readonly description: "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.";
50
+ readonly created_at: "2024-04-18 20:31:47.273";
47
51
  readonly tags: readonly [];
48
52
  readonly properties: {
53
+ readonly price: readonly [{
54
+ readonly unit: "per M input tokens";
55
+ readonly price: "$0.28";
56
+ }, {
57
+ readonly unit: "per M output tokens";
58
+ readonly price: "$0.83";
59
+ }];
49
60
  readonly context_window: 7968;
50
61
  readonly info: "https://llama.meta.com";
51
62
  readonly terms: "https://llama.meta.com/llama3/license/#";
@@ -55,8 +66,16 @@ export declare const workersAiCatalog: {
55
66
  readonly source: 1;
56
67
  readonly name: "@cf/meta/llama-3.2-3b-instruct";
57
68
  readonly description: "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
69
+ readonly created_at: "2024-09-25 20:05:43.986";
58
70
  readonly tags: readonly [];
59
71
  readonly properties: {
72
+ readonly price: readonly [{
73
+ readonly unit: "per M input tokens";
74
+ readonly price: "$0.051";
75
+ }, {
76
+ readonly unit: "per M output tokens";
77
+ readonly price: "$0.34";
78
+ }];
60
79
  readonly context_window: 128000;
61
80
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
62
81
  };
@@ -65,6 +84,7 @@ export declare const workersAiCatalog: {
65
84
  readonly source: 2;
66
85
  readonly name: "@hf/thebloke/llamaguard-7b-awq";
67
86
  readonly description: "Llama Guard is a model for classifying the safety of LLM prompts and responses, using a taxonomy of safety risks.\n";
87
+ readonly created_at: "2024-02-06 18:13:59.060";
68
88
  readonly tags: readonly [];
69
89
  readonly properties: {
70
90
  readonly beta: true;
@@ -75,6 +95,7 @@ export declare const workersAiCatalog: {
75
95
  readonly source: 2;
76
96
  readonly name: "@hf/thebloke/neural-chat-7b-v3-1-awq";
77
97
  readonly description: "This model is a fine-tuned 7B parameter LLM on the Intel Gaudi 2 processor from the mistralai/Mistral-7B-v0.1 on the open source dataset Open-Orca/SlimOrca.";
98
+ readonly created_at: "2024-02-06 18:12:30.722";
78
99
  readonly tags: readonly [];
79
100
  readonly properties: {
80
101
  readonly beta: true;
@@ -85,13 +106,23 @@ export declare const workersAiCatalog: {
85
106
  readonly source: 1;
86
107
  readonly name: "@cf/meta/llama-guard-3-8b";
87
108
  readonly description: "Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.";
109
+ readonly created_at: "2025-01-22 23:26:23.495";
88
110
  readonly tags: readonly [];
89
- readonly properties: {};
111
+ readonly properties: {
112
+ readonly price: readonly [{
113
+ readonly unit: "per M input tokens";
114
+ readonly price: "$0.48";
115
+ }, {
116
+ readonly unit: "per M output tokens";
117
+ readonly price: "$0.030";
118
+ }];
119
+ };
90
120
  }, {
91
121
  readonly id: "ca54bcd6-0d98-4739-9b3b-5c8b4402193d";
92
122
  readonly source: 1;
93
123
  readonly name: "@cf/meta/llama-2-7b-chat-fp16";
94
124
  readonly description: "Full precision (fp16) generative text model with 7 billion parameters from Meta";
125
+ readonly created_at: "2023-11-07 11:54:20.229";
95
126
  readonly tags: readonly [];
96
127
  readonly properties: {
97
128
  readonly beta: true;
@@ -104,6 +135,7 @@ export declare const workersAiCatalog: {
104
135
  readonly source: 1;
105
136
  readonly name: "@cf/mistral/mistral-7b-instruct-v0.1";
106
137
  readonly description: "Instruct fine-tuned version of the Mistral-7b generative text model with 7 billion parameters";
138
+ readonly created_at: "2023-11-07 11:54:20.229";
107
139
  readonly tags: readonly [];
108
140
  readonly properties: {
109
141
  readonly beta: true;
@@ -116,6 +148,7 @@ export declare const workersAiCatalog: {
116
148
  readonly source: 1;
117
149
  readonly name: "@cf/mistral/mistral-7b-instruct-v0.2-lora";
118
150
  readonly description: "The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2.";
151
+ readonly created_at: "2024-04-01 22:14:40.529";
119
152
  readonly tags: readonly [];
120
153
  readonly properties: {
121
154
  readonly beta: true;
@@ -127,6 +160,7 @@ export declare const workersAiCatalog: {
127
160
  readonly source: 1;
128
161
  readonly name: "@cf/tinyllama/tinyllama-1.1b-chat-v1.0";
129
162
  readonly description: "The TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens. This is the chat model finetuned on top of TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T.";
163
+ readonly created_at: "2024-02-27 18:25:37.524";
130
164
  readonly tags: readonly [];
131
165
  readonly properties: {
132
166
  readonly beta: true;
@@ -138,6 +172,7 @@ export declare const workersAiCatalog: {
138
172
  readonly source: 2;
139
173
  readonly name: "@hf/mistral/mistral-7b-instruct-v0.2";
140
174
  readonly description: "The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2. Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1: 32k context window (vs 8k context in v0.1), rope-theta = 1e6, and no Sliding-Window Attention.";
175
+ readonly created_at: "2024-04-02 13:00:59.244";
141
176
  readonly tags: readonly [];
142
177
  readonly properties: {
143
178
  readonly beta: true;
@@ -153,6 +188,7 @@ export declare const workersAiCatalog: {
153
188
  readonly source: 1;
154
189
  readonly name: "@cf/fblgit/una-cybertron-7b-v2-bf16";
155
190
  readonly description: "Cybertron 7B v2 is a 7B MistralAI based model, best on it's series. It was trained with SFT, DPO and UNA (Unified Neural Alignment) on multiple datasets.";
191
+ readonly created_at: "2024-04-24 14:37:19.494";
156
192
  readonly tags: readonly [];
157
193
  readonly properties: {
158
194
  readonly beta: true;
@@ -163,9 +199,16 @@ export declare const workersAiCatalog: {
163
199
  readonly source: 1;
164
200
  readonly name: "@cf/deepseek-ai/deepseek-r1-distill-qwen-32b";
165
201
  readonly description: "DeepSeek-R1-Distill-Qwen-32B is a model distilled from DeepSeek-R1 based on Qwen2.5. It outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.";
202
+ readonly created_at: "2025-01-22 19:48:55.776";
166
203
  readonly tags: readonly [];
167
204
  readonly properties: {
168
- readonly context_window: 80000;
205
+ readonly price: readonly [{
206
+ readonly unit: "per M input tokens";
207
+ readonly price: "$0.50";
208
+ }, {
209
+ readonly unit: "per M output tokens";
210
+ readonly price: "$4.88";
211
+ }];
169
212
  readonly terms: "https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE";
170
213
  };
171
214
  }, {
@@ -173,6 +216,7 @@ export declare const workersAiCatalog: {
173
216
  readonly source: 1;
174
217
  readonly name: "@cf/thebloke/discolm-german-7b-v1-awq";
175
218
  readonly description: "DiscoLM German 7b is a Mistral-based large language model with a focus on German-language applications. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.";
219
+ readonly created_at: "2024-02-27 18:23:05.178";
176
220
  readonly tags: readonly [];
177
221
  readonly properties: {
178
222
  readonly beta: true;
@@ -184,6 +228,7 @@ export declare const workersAiCatalog: {
184
228
  readonly source: 1;
185
229
  readonly name: "@cf/meta/llama-2-7b-chat-int8";
186
230
  readonly description: "Quantized (int8) generative text model with 7 billion parameters from Meta";
231
+ readonly created_at: "2023-09-25 19:21:11.898";
187
232
  readonly tags: readonly [];
188
233
  readonly properties: {
189
234
  readonly context_window: 8192;
@@ -193,8 +238,16 @@ export declare const workersAiCatalog: {
193
238
  readonly source: 1;
194
239
  readonly name: "@cf/meta/llama-3.1-8b-instruct-fp8";
195
240
  readonly description: "Llama 3.1 8B quantized to FP8 precision";
241
+ readonly created_at: "2024-07-25 17:28:43.328";
196
242
  readonly tags: readonly [];
197
243
  readonly properties: {
244
+ readonly price: readonly [{
245
+ readonly unit: "per M input tokens";
246
+ readonly price: "$0.15";
247
+ }, {
248
+ readonly unit: "per M output tokens";
249
+ readonly price: "$0.29";
250
+ }];
198
251
  readonly context_window: 32000;
199
252
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
200
253
  };
@@ -203,6 +256,7 @@ export declare const workersAiCatalog: {
203
256
  readonly source: 2;
204
257
  readonly name: "@hf/thebloke/mistral-7b-instruct-v0.1-awq";
205
258
  readonly description: "Mistral 7B Instruct v0.1 AWQ is an efficient, accurate and blazing-fast low-bit weight quantized Mistral variant.";
259
+ readonly created_at: "2023-11-24 00:27:15.869";
206
260
  readonly tags: readonly [];
207
261
  readonly properties: {
208
262
  readonly beta: true;
@@ -214,6 +268,7 @@ export declare const workersAiCatalog: {
214
268
  readonly source: 1;
215
269
  readonly name: "@cf/qwen/qwen1.5-7b-chat-awq";
216
270
  readonly description: "Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.";
271
+ readonly created_at: "2024-02-27 18:24:11.709";
217
272
  readonly tags: readonly [];
218
273
  readonly properties: {
219
274
  readonly beta: true;
@@ -225,8 +280,16 @@ export declare const workersAiCatalog: {
225
280
  readonly source: 1;
226
281
  readonly name: "@cf/meta/llama-3.2-1b-instruct";
227
282
  readonly description: "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
283
+ readonly created_at: "2024-09-25 21:36:32.050";
228
284
  readonly tags: readonly [];
229
285
  readonly properties: {
286
+ readonly price: readonly [{
287
+ readonly unit: "per M input tokens";
288
+ readonly price: "$0.027";
289
+ }, {
290
+ readonly unit: "per M output tokens";
291
+ readonly price: "$0.20";
292
+ }];
230
293
  readonly context_window: 60000;
231
294
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
232
295
  };
@@ -235,6 +298,7 @@ export declare const workersAiCatalog: {
235
298
  readonly source: 2;
236
299
  readonly name: "@hf/thebloke/llama-2-13b-chat-awq";
237
300
  readonly description: "Llama 2 13B Chat AWQ is an efficient, accurate and blazing-fast low-bit weight quantized Llama 2 variant.";
301
+ readonly created_at: "2023-11-24 00:27:15.869";
238
302
  readonly tags: readonly [];
239
303
  readonly properties: {
240
304
  readonly beta: true;
@@ -246,6 +310,7 @@ export declare const workersAiCatalog: {
246
310
  readonly source: 2;
247
311
  readonly name: "@hf/thebloke/deepseek-coder-6.7b-base-awq";
248
312
  readonly description: "Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese.";
313
+ readonly created_at: "2024-02-06 18:16:27.183";
249
314
  readonly tags: readonly [];
250
315
  readonly properties: {
251
316
  readonly beta: true;
@@ -257,6 +322,7 @@ export declare const workersAiCatalog: {
257
322
  readonly source: 1;
258
323
  readonly name: "@cf/meta-llama/llama-2-7b-chat-hf-lora";
259
324
  readonly description: "This is a Llama2 base model that Cloudflare dedicated for inference with LoRA adapters. Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 7B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format. ";
325
+ readonly created_at: "2024-04-02 00:17:18.579";
260
326
  readonly tags: readonly [];
261
327
  readonly properties: {
262
328
  readonly beta: true;
@@ -268,8 +334,16 @@ export declare const workersAiCatalog: {
268
334
  readonly source: 1;
269
335
  readonly name: "@cf/meta/llama-3.3-70b-instruct-fp8-fast";
270
336
  readonly description: "Llama 3.3 70B quantized to fp8 precision, optimized to be faster.";
337
+ readonly created_at: "2024-12-06 17:09:18.338";
271
338
  readonly tags: readonly [];
272
339
  readonly properties: {
340
+ readonly price: readonly [{
341
+ readonly unit: "per M input tokens";
342
+ readonly price: "$0.29";
343
+ }, {
344
+ readonly unit: "per M output tokens";
345
+ readonly price: "$2.25";
346
+ }];
273
347
  readonly context_window: 24000;
274
348
  readonly function_calling: true;
275
349
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_3/LICENSE";
@@ -279,6 +353,7 @@ export declare const workersAiCatalog: {
279
353
  readonly source: 2;
280
354
  readonly name: "@hf/thebloke/openhermes-2.5-mistral-7b-awq";
281
355
  readonly description: "OpenHermes 2.5 Mistral 7B is a state of the art Mistral Fine-tune, a continuation of OpenHermes 2 model, which trained on additional code datasets.";
356
+ readonly created_at: "2024-02-06 18:04:22.846";
282
357
  readonly tags: readonly [];
283
358
  readonly properties: {
284
359
  readonly beta: true;
@@ -289,6 +364,7 @@ export declare const workersAiCatalog: {
289
364
  readonly source: 2;
290
365
  readonly name: "@hf/thebloke/deepseek-coder-6.7b-instruct-awq";
291
366
  readonly description: "Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese.";
367
+ readonly created_at: "2024-02-06 18:18:27.462";
292
368
  readonly tags: readonly [];
293
369
  readonly properties: {
294
370
  readonly beta: true;
@@ -300,6 +376,7 @@ export declare const workersAiCatalog: {
300
376
  readonly source: 1;
301
377
  readonly name: "@cf/deepseek-ai/deepseek-math-7b-instruct";
302
378
  readonly description: "DeepSeekMath-Instruct 7B is a mathematically instructed tuning model derived from DeepSeekMath-Base 7B. DeepSeekMath is initialized with DeepSeek-Coder-v1.5 7B and continues pre-training on math-related tokens sourced from Common Crawl, together with natural language and code data for 500B tokens.";
379
+ readonly created_at: "2024-02-27 17:54:17.459";
303
380
  readonly tags: readonly [];
304
381
  readonly properties: {
305
382
  readonly beta: true;
@@ -312,6 +389,7 @@ export declare const workersAiCatalog: {
312
389
  readonly source: 1;
313
390
  readonly name: "@cf/tiiuae/falcon-7b-instruct";
314
391
  readonly description: "Falcon-7B-Instruct is a 7B parameters causal decoder-only model built by TII based on Falcon-7B and finetuned on a mixture of chat/instruct datasets.";
392
+ readonly created_at: "2024-02-27 18:21:15.796";
315
393
  readonly tags: readonly [];
316
394
  readonly properties: {
317
395
  readonly beta: true;
@@ -323,6 +401,7 @@ export declare const workersAiCatalog: {
323
401
  readonly source: 2;
324
402
  readonly name: "@hf/nousresearch/hermes-2-pro-mistral-7b";
325
403
  readonly description: "Hermes 2 Pro on Mistral 7B is the new flagship 7B Hermes! Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.";
404
+ readonly created_at: "2024-04-01 23:45:53.800";
326
405
  readonly tags: readonly [];
327
406
  readonly properties: {
328
407
  readonly beta: true;
@@ -335,8 +414,16 @@ export declare const workersAiCatalog: {
335
414
  readonly source: 1;
336
415
  readonly name: "@cf/meta/llama-3.1-8b-instruct";
337
416
  readonly description: "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.";
417
+ readonly created_at: "2024-07-18 22:53:33.746";
338
418
  readonly tags: readonly [];
339
419
  readonly properties: {
420
+ readonly price: readonly [{
421
+ readonly unit: "per M input tokens";
422
+ readonly price: "$0.28";
423
+ }, {
424
+ readonly unit: "per M output tokens";
425
+ readonly price: "$0.83";
426
+ }];
340
427
  readonly context_window: 7968;
341
428
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
342
429
  };
@@ -345,8 +432,16 @@ export declare const workersAiCatalog: {
345
432
  readonly source: 1;
346
433
  readonly name: "@cf/meta/llama-3.1-8b-instruct-awq";
347
434
  readonly description: "Quantized (int4) generative text model with 8 billion parameters from Meta.\n";
435
+ readonly created_at: "2024-07-25 17:46:04.304";
348
436
  readonly tags: readonly [];
349
437
  readonly properties: {
438
+ readonly price: readonly [{
439
+ readonly unit: "per M input tokens";
440
+ readonly price: "$0.12";
441
+ }, {
442
+ readonly unit: "per M output tokens";
443
+ readonly price: "$0.27";
444
+ }];
350
445
  readonly context_window: 8192;
351
446
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
352
447
  };
@@ -355,6 +450,7 @@ export declare const workersAiCatalog: {
355
450
  readonly source: 2;
356
451
  readonly name: "@hf/thebloke/zephyr-7b-beta-awq";
357
452
  readonly description: "Zephyr 7B Beta AWQ is an efficient, accurate and blazing-fast low-bit weight quantized Zephyr model variant.";
453
+ readonly created_at: "2023-11-24 00:27:15.869";
358
454
  readonly tags: readonly [];
359
455
  readonly properties: {
360
456
  readonly beta: true;
@@ -366,6 +462,7 @@ export declare const workersAiCatalog: {
366
462
  readonly source: 1;
367
463
  readonly name: "@cf/google/gemma-7b-it-lora";
368
464
  readonly description: " This is a Gemma-7B base model that Cloudflare dedicates for inference with LoRA adapters. Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.";
465
+ readonly created_at: "2024-04-02 00:20:19.633";
369
466
  readonly tags: readonly [];
370
467
  readonly properties: {
371
468
  readonly beta: true;
@@ -377,6 +474,7 @@ export declare const workersAiCatalog: {
377
474
  readonly source: 1;
378
475
  readonly name: "@cf/qwen/qwen1.5-1.8b-chat";
379
476
  readonly description: "Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud.";
477
+ readonly created_at: "2024-02-27 18:30:31.723";
380
478
  readonly tags: readonly [];
381
479
  readonly properties: {
382
480
  readonly beta: true;
@@ -388,8 +486,16 @@ export declare const workersAiCatalog: {
388
486
  readonly source: 1;
389
487
  readonly name: "@cf/meta/llama-3-8b-instruct-awq";
390
488
  readonly description: "Quantized (int4) generative text model with 8 billion parameters from Meta.";
489
+ readonly created_at: "2024-05-09 23:32:47.584";
391
490
  readonly tags: readonly [];
392
491
  readonly properties: {
492
+ readonly price: readonly [{
493
+ readonly unit: "per M input tokens";
494
+ readonly price: "$0.12";
495
+ }, {
496
+ readonly unit: "per M output tokens";
497
+ readonly price: "$0.27";
498
+ }];
393
499
  readonly context_window: 8192;
394
500
  readonly info: "https://llama.meta.com";
395
501
  readonly terms: "https://llama.meta.com/llama3/license/#";
@@ -399,8 +505,16 @@ export declare const workersAiCatalog: {
399
505
  readonly source: 1;
400
506
  readonly name: "@cf/meta/llama-3.2-11b-vision-instruct";
401
507
  readonly description: " The Llama 3.2-Vision instruction-tuned models are optimized for visual recognition, image reasoning, captioning, and answering general questions about an image.";
508
+ readonly created_at: "2024-09-25 05:36:04.547";
402
509
  readonly tags: readonly [];
403
510
  readonly properties: {
511
+ readonly price: readonly [{
512
+ readonly unit: "per M input tokens";
513
+ readonly price: "$0.049";
514
+ }, {
515
+ readonly unit: "per M output tokens";
516
+ readonly price: "$0.68";
517
+ }];
404
518
  readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
405
519
  };
406
520
  }, {
@@ -408,6 +522,7 @@ export declare const workersAiCatalog: {
408
522
  readonly source: 1;
409
523
  readonly name: "@cf/defog/sqlcoder-7b-2";
410
524
  readonly description: "This model is intended to be used by non-technical users to understand data inside their SQL databases. ";
525
+ readonly created_at: "2024-02-27 18:18:46.095";
411
526
  readonly tags: readonly [];
412
527
  readonly properties: {
413
528
  readonly beta: true;
@@ -420,6 +535,7 @@ export declare const workersAiCatalog: {
420
535
  readonly source: 1;
421
536
  readonly name: "@cf/microsoft/phi-2";
422
537
  readonly description: "Phi-2 is a Transformer-based model with a next-word prediction objective, trained on 1.4T tokens from multiple passes on a mixture of Synthetic and Web datasets for NLP and coding.";
538
+ readonly created_at: "2024-02-27 18:26:21.126";
423
539
  readonly tags: readonly [];
424
540
  readonly properties: {
425
541
  readonly beta: true;
@@ -431,6 +547,7 @@ export declare const workersAiCatalog: {
431
547
  readonly source: 2;
432
548
  readonly name: "@hf/meta-llama/meta-llama-3-8b-instruct";
433
549
  readonly description: "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.\t";
550
+ readonly created_at: "2024-05-22 18:21:04.371";
434
551
  readonly tags: readonly [];
435
552
  readonly properties: {
436
553
  readonly context_window: 8192;
@@ -440,6 +557,7 @@ export declare const workersAiCatalog: {
440
557
  readonly source: 2;
441
558
  readonly name: "@hf/google/gemma-7b-it";
442
559
  readonly description: "Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants.";
560
+ readonly created_at: "2024-04-01 23:51:35.866";
443
561
  readonly tags: readonly [];
444
562
  readonly properties: {
445
563
  readonly beta: true;
@@ -453,6 +571,7 @@ export declare const workersAiCatalog: {
453
571
  readonly source: 1;
454
572
  readonly name: "@cf/qwen/qwen1.5-14b-chat-awq";
455
573
  readonly description: "Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.";
574
+ readonly created_at: "2024-02-27 18:24:45.316";
456
575
  readonly tags: readonly [];
457
576
  readonly properties: {
458
577
  readonly beta: true;
@@ -464,6 +583,7 @@ export declare const workersAiCatalog: {
464
583
  readonly source: 1;
465
584
  readonly name: "@cf/openchat/openchat-3.5-0106";
466
585
  readonly description: "OpenChat is an innovative library of open-source language models, fine-tuned with C-RLFT - a strategy inspired by offline reinforcement learning.";
586
+ readonly created_at: "2024-02-27 18:20:39.169";
467
587
  readonly tags: readonly [];
468
588
  readonly properties: {
469
589
  readonly beta: true;
@@ -480,6 +600,7 @@ export declare const workersAiCatalog: {
480
600
  readonly source: 1;
481
601
  readonly name: "@cf/baai/bge-m3";
482
602
  readonly description: "Multi-Functionality, Multi-Linguality, and Multi-Granularity embeddings model.";
603
+ readonly created_at: "2024-05-22 19:27:09.781";
483
604
  readonly tags: readonly [];
484
605
  readonly properties: {};
485
606
  }, {
@@ -487,6 +608,7 @@ export declare const workersAiCatalog: {
487
608
  readonly source: 1;
488
609
  readonly name: "@cf/baai/bge-small-en-v1.5";
489
610
  readonly description: "BAAI general embedding (Small) model that transforms any given text into a 384-dimensional vector";
611
+ readonly created_at: "2023-11-07 15:43:58.042";
490
612
  readonly tags: readonly [];
491
613
  readonly properties: {
492
614
  readonly beta: true;
@@ -499,6 +621,7 @@ export declare const workersAiCatalog: {
499
621
  readonly source: 1;
500
622
  readonly name: "@cf/baai/bge-base-en-v1.5";
501
623
  readonly description: "BAAI general embedding (Base) model that transforms any given text into a 768-dimensional vector";
624
+ readonly created_at: "2023-09-25 19:21:11.898";
502
625
  readonly tags: readonly [];
503
626
  readonly properties: {
504
627
  readonly beta: true;
@@ -511,6 +634,7 @@ export declare const workersAiCatalog: {
511
634
  readonly source: 1;
512
635
  readonly name: "@cf/baai/bge-large-en-v1.5";
513
636
  readonly description: "BAAI general embedding (Large) model that transforms any given text into a 1024-dimensional vector";
637
+ readonly created_at: "2023-11-07 15:43:58.042";
514
638
  readonly tags: readonly [];
515
639
  readonly properties: {
516
640
  readonly beta: true;
@@ -528,6 +652,7 @@ export declare const workersAiCatalog: {
528
652
  readonly source: 1;
529
653
  readonly name: "@cf/huggingface/distilbert-sst-2-int8";
530
654
  readonly description: "Distilled BERT model that was finetuned on SST-2 for sentiment classification";
655
+ readonly created_at: "2023-09-25 19:21:11.898";
531
656
  readonly tags: readonly [];
532
657
  readonly properties: {
533
658
  readonly beta: true;
@@ -538,6 +663,7 @@ export declare const workersAiCatalog: {
538
663
  readonly source: 1;
539
664
  readonly name: "@cf/baai/bge-reranker-base";
540
665
  readonly description: "Different from embedding model, reranker uses question and document as input and directly output similarity instead of embedding. You can get a relevance score by inputting query and passage to the reranker. And the score can be mapped to a float value in [0,1] by sigmoid function.\n\n";
666
+ readonly created_at: "2025-02-14 12:28:19.009";
541
667
  readonly tags: readonly [];
542
668
  readonly properties: {};
543
669
  }];
@@ -550,6 +676,7 @@ export declare const workersAiCatalog: {
550
676
  readonly source: 1;
551
677
  readonly name: "@cf/facebook/detr-resnet-50";
552
678
  readonly description: "DEtection TRansformer (DETR) model trained end-to-end on COCO 2017 object detection (118k annotated images).";
679
+ readonly created_at: "2024-02-27 17:43:51.922";
553
680
  readonly tags: readonly [];
554
681
  readonly properties: {
555
682
  readonly beta: true;
@@ -564,6 +691,7 @@ export declare const workersAiCatalog: {
564
691
  readonly source: 1;
565
692
  readonly name: "@cf/myshell-ai/melotts";
566
693
  readonly description: "MeloTTS is a high-quality multi-lingual text-to-speech library by MyShell.ai.";
694
+ readonly created_at: "2024-07-19 15:51:04.819";
567
695
  readonly tags: readonly [];
568
696
  readonly properties: {};
569
697
  }];
@@ -576,6 +704,7 @@ export declare const workersAiCatalog: {
576
704
  readonly source: 1;
577
705
  readonly name: "@cf/openai/whisper";
578
706
  readonly description: "Whisper is a general-purpose speech recognition model. It is trained on a large dataset of diverse audio and is also a multitasking model that can perform multilingual speech recognition, speech translation, and language identification.";
707
+ readonly created_at: "2023-09-25 19:21:11.898";
579
708
  readonly tags: readonly [];
580
709
  readonly properties: {
581
710
  readonly beta: true;
@@ -586,6 +715,7 @@ export declare const workersAiCatalog: {
586
715
  readonly source: 1;
587
716
  readonly name: "@cf/openai/whisper-tiny-en";
588
717
  readonly description: "Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours of labelled data, Whisper models demonstrate a strong ability to generalize to many datasets and domains without the need for fine-tuning. This is the English-only version of the Whisper Tiny model which was trained on the task of speech recognition.";
718
+ readonly created_at: "2024-04-22 20:59:02.731";
589
719
  readonly tags: readonly [];
590
720
  readonly properties: {
591
721
  readonly beta: true;
@@ -595,8 +725,14 @@ export declare const workersAiCatalog: {
595
725
  readonly source: 1;
596
726
  readonly name: "@cf/openai/whisper-large-v3-turbo";
597
727
  readonly description: "Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. ";
728
+ readonly created_at: "2024-05-22 00:02:18.656";
598
729
  readonly tags: readonly [];
599
- readonly properties: {};
730
+ readonly properties: {
731
+ readonly price: readonly [{
732
+ readonly unit: "per audio minute";
733
+ readonly price: "$0.00051";
734
+ }];
735
+ };
600
736
  }];
601
737
  };
602
738
  readonly 'Image-to-Text': {
@@ -607,6 +743,7 @@ export declare const workersAiCatalog: {
607
743
  readonly source: 1;
608
744
  readonly name: "@cf/llava-hf/llava-1.5-7b-hf";
609
745
  readonly description: "LLaVA is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-generated multimodal instruction-following data. It is an auto-regressive language model, based on the transformer architecture.";
746
+ readonly created_at: "2024-05-01 18:00:39.971";
610
747
  readonly tags: readonly [];
611
748
  readonly properties: {
612
749
  readonly beta: true;
@@ -616,6 +753,7 @@ export declare const workersAiCatalog: {
616
753
  readonly source: 1;
617
754
  readonly name: "@cf/unum/uform-gen2-qwen-500m";
618
755
  readonly description: "UForm-Gen is a small generative vision-language model primarily designed for Image Captioning and Visual Question Answering. The model was pre-trained on the internal image captioning dataset and fine-tuned on public instructions datasets: SVIT, LVIS, VQAs datasets.";
756
+ readonly created_at: "2024-02-27 18:28:52.485";
619
757
  readonly tags: readonly [];
620
758
  readonly properties: {
621
759
  readonly beta: true;
@@ -631,6 +769,7 @@ export declare const workersAiCatalog: {
631
769
  readonly source: 1;
632
770
  readonly name: "@cf/runwayml/stable-diffusion-v1-5-inpainting";
633
771
  readonly description: "Stable Diffusion Inpainting is a latent text-to-image diffusion model capable of generating photo-realistic images given any text input, with the extra capability of inpainting the pictures by using a mask.";
772
+ readonly created_at: "2024-02-27 17:23:57.528";
634
773
  readonly tags: readonly [];
635
774
  readonly properties: {
636
775
  readonly beta: true;
@@ -642,6 +781,7 @@ export declare const workersAiCatalog: {
642
781
  readonly source: 1;
643
782
  readonly name: "@cf/black-forest-labs/flux-1-schnell";
644
783
  readonly description: "FLUX.1 [schnell] is a 12 billion parameter rectified flow transformer capable of generating images from text descriptions. ";
784
+ readonly created_at: "2024-08-29 16:37:39.541";
645
785
  readonly tags: readonly [];
646
786
  readonly properties: {};
647
787
  }, {
@@ -649,6 +789,7 @@ export declare const workersAiCatalog: {
649
789
  readonly source: 1;
650
790
  readonly name: "@cf/bytedance/stable-diffusion-xl-lightning";
651
791
  readonly description: "SDXL-Lightning is a lightning-fast text-to-image generation model. It can generate high-quality 1024px images in a few steps.";
792
+ readonly created_at: "2024-02-27 17:41:29.578";
652
793
  readonly tags: readonly [];
653
794
  readonly properties: {
654
795
  readonly beta: true;
@@ -659,6 +800,7 @@ export declare const workersAiCatalog: {
659
800
  readonly source: 1;
660
801
  readonly name: "@cf/lykon/dreamshaper-8-lcm";
661
802
  readonly description: "Stable Diffusion model that has been fine-tuned to be better at photorealism without sacrificing range.";
803
+ readonly created_at: "2024-02-27 17:40:38.881";
662
804
  readonly tags: readonly [];
663
805
  readonly properties: {
664
806
  readonly beta: true;
@@ -669,6 +811,7 @@ export declare const workersAiCatalog: {
669
811
  readonly source: 1;
670
812
  readonly name: "@cf/stabilityai/stable-diffusion-xl-base-1.0";
671
813
  readonly description: "Diffusion-based text-to-image generative model by Stability AI. Generates and modify images based on text prompts.";
814
+ readonly created_at: "2023-11-10 10:54:43.694";
672
815
  readonly tags: readonly [];
673
816
  readonly properties: {
674
817
  readonly beta: true;
@@ -680,6 +823,7 @@ export declare const workersAiCatalog: {
680
823
  readonly source: 1;
681
824
  readonly name: "@cf/runwayml/stable-diffusion-v1-5-img2img";
682
825
  readonly description: "Stable Diffusion is a latent text-to-image diffusion model capable of generating photo-realistic images. Img2img generate a new image from an input image with Stable Diffusion. ";
826
+ readonly created_at: "2024-02-27 17:32:28.581";
683
827
  readonly tags: readonly [];
684
828
  readonly properties: {
685
829
  readonly beta: true;
@@ -696,6 +840,7 @@ export declare const workersAiCatalog: {
696
840
  readonly source: 1;
697
841
  readonly name: "@cf/microsoft/resnet-50";
698
842
  readonly description: "50 layers deep image classification CNN trained on more than 1M images from ImageNet";
843
+ readonly created_at: "2023-09-25 19:21:11.898";
699
844
  readonly tags: readonly [];
700
845
  readonly properties: {
701
846
  readonly beta: true;
@@ -711,6 +856,7 @@ export declare const workersAiCatalog: {
711
856
  readonly source: 1;
712
857
  readonly name: "@cf/meta/m2m100-1.2b";
713
858
  readonly description: "Multilingual encoder-decoder (seq-to-seq) model trained for Many-to-Many multilingual translation";
859
+ readonly created_at: "2023-09-25 19:21:11.898";
714
860
  readonly tags: readonly [];
715
861
  readonly properties: {
716
862
  readonly beta: true;
@@ -728,6 +874,7 @@ export declare const workersAiCatalog: {
728
874
  readonly source: 1;
729
875
  readonly name: "@cf/facebook/bart-large-cnn";
730
876
  readonly description: "BART is a transformer encoder-encoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. You can use this model for text summarization.";
877
+ readonly created_at: "2024-02-27 18:28:11.833";
731
878
  readonly tags: readonly [];
732
879
  readonly properties: {
733
880
  readonly beta: true;