@chainfuse/types 2.2.0 → 2.2.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/ai-tools/azure/catalog.d.ts +239 -237
- package/dist/ai-tools/azure/catalog.js +267 -265
- package/dist/ai-tools/workers-ai/catalog.d.ts +150 -3
- package/dist/ai-tools/workers-ai/catalog.js +185 -3
- package/dist/wf/index.js +1 -1
- package/package.json +6 -6
|
@@ -8,6 +8,7 @@ export declare const workersAiCatalog: {
|
|
|
8
8
|
readonly source: 1;
|
|
9
9
|
readonly name: "@cf/qwen/qwen1.5-0.5b-chat";
|
|
10
10
|
readonly description: "Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud.";
|
|
11
|
+
readonly created_at: "2024-02-27 18:23:37.344";
|
|
11
12
|
readonly tags: readonly [];
|
|
12
13
|
readonly properties: {
|
|
13
14
|
readonly beta: true;
|
|
@@ -19,6 +20,7 @@ export declare const workersAiCatalog: {
|
|
|
19
20
|
readonly source: 1;
|
|
20
21
|
readonly name: "@cf/google/gemma-2b-it-lora";
|
|
21
22
|
readonly description: "This is a Gemma-2B base model that Cloudflare dedicates for inference with LoRA adapters. Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.";
|
|
23
|
+
readonly created_at: "2024-04-02 00:19:34.669";
|
|
22
24
|
readonly tags: readonly [];
|
|
23
25
|
readonly properties: {
|
|
24
26
|
readonly beta: true;
|
|
@@ -30,6 +32,7 @@ export declare const workersAiCatalog: {
|
|
|
30
32
|
readonly source: 2;
|
|
31
33
|
readonly name: "@hf/nexusflow/starling-lm-7b-beta";
|
|
32
34
|
readonly description: "We introduce Starling-LM-7B-beta, an open large language model (LLM) trained by Reinforcement Learning from AI Feedback (RLAIF). Starling-LM-7B-beta is trained from Openchat-3.5-0106 with our new reward model Nexusflow/Starling-RM-34B and policy optimization method Fine-Tuning Language Models from Human Preferences (PPO).";
|
|
35
|
+
readonly created_at: "2024-04-01 23:49:31.797";
|
|
33
36
|
readonly tags: readonly [];
|
|
34
37
|
readonly properties: {
|
|
35
38
|
readonly beta: true;
|
|
@@ -44,8 +47,16 @@ export declare const workersAiCatalog: {
|
|
|
44
47
|
readonly source: 1;
|
|
45
48
|
readonly name: "@cf/meta/llama-3-8b-instruct";
|
|
46
49
|
readonly description: "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.";
|
|
50
|
+
readonly created_at: "2024-04-18 20:31:47.273";
|
|
47
51
|
readonly tags: readonly [];
|
|
48
52
|
readonly properties: {
|
|
53
|
+
readonly price: readonly [{
|
|
54
|
+
readonly unit: "per M input tokens";
|
|
55
|
+
readonly price: "$0.28";
|
|
56
|
+
}, {
|
|
57
|
+
readonly unit: "per M output tokens";
|
|
58
|
+
readonly price: "$0.83";
|
|
59
|
+
}];
|
|
49
60
|
readonly context_window: 7968;
|
|
50
61
|
readonly info: "https://llama.meta.com";
|
|
51
62
|
readonly terms: "https://llama.meta.com/llama3/license/#";
|
|
@@ -55,8 +66,16 @@ export declare const workersAiCatalog: {
|
|
|
55
66
|
readonly source: 1;
|
|
56
67
|
readonly name: "@cf/meta/llama-3.2-3b-instruct";
|
|
57
68
|
readonly description: "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
|
|
69
|
+
readonly created_at: "2024-09-25 20:05:43.986";
|
|
58
70
|
readonly tags: readonly [];
|
|
59
71
|
readonly properties: {
|
|
72
|
+
readonly price: readonly [{
|
|
73
|
+
readonly unit: "per M input tokens";
|
|
74
|
+
readonly price: "$0.051";
|
|
75
|
+
}, {
|
|
76
|
+
readonly unit: "per M output tokens";
|
|
77
|
+
readonly price: "$0.34";
|
|
78
|
+
}];
|
|
60
79
|
readonly context_window: 128000;
|
|
61
80
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
|
|
62
81
|
};
|
|
@@ -65,6 +84,7 @@ export declare const workersAiCatalog: {
|
|
|
65
84
|
readonly source: 2;
|
|
66
85
|
readonly name: "@hf/thebloke/llamaguard-7b-awq";
|
|
67
86
|
readonly description: "Llama Guard is a model for classifying the safety of LLM prompts and responses, using a taxonomy of safety risks.\n";
|
|
87
|
+
readonly created_at: "2024-02-06 18:13:59.060";
|
|
68
88
|
readonly tags: readonly [];
|
|
69
89
|
readonly properties: {
|
|
70
90
|
readonly beta: true;
|
|
@@ -75,6 +95,7 @@ export declare const workersAiCatalog: {
|
|
|
75
95
|
readonly source: 2;
|
|
76
96
|
readonly name: "@hf/thebloke/neural-chat-7b-v3-1-awq";
|
|
77
97
|
readonly description: "This model is a fine-tuned 7B parameter LLM on the Intel Gaudi 2 processor from the mistralai/Mistral-7B-v0.1 on the open source dataset Open-Orca/SlimOrca.";
|
|
98
|
+
readonly created_at: "2024-02-06 18:12:30.722";
|
|
78
99
|
readonly tags: readonly [];
|
|
79
100
|
readonly properties: {
|
|
80
101
|
readonly beta: true;
|
|
@@ -85,13 +106,23 @@ export declare const workersAiCatalog: {
|
|
|
85
106
|
readonly source: 1;
|
|
86
107
|
readonly name: "@cf/meta/llama-guard-3-8b";
|
|
87
108
|
readonly description: "Llama Guard 3 is a Llama-3.1-8B pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM – it generates text in its output that indicates whether a given prompt or response is safe or unsafe, and if unsafe, it also lists the content categories violated.";
|
|
109
|
+
readonly created_at: "2025-01-22 23:26:23.495";
|
|
88
110
|
readonly tags: readonly [];
|
|
89
|
-
readonly properties: {
|
|
111
|
+
readonly properties: {
|
|
112
|
+
readonly price: readonly [{
|
|
113
|
+
readonly unit: "per M input tokens";
|
|
114
|
+
readonly price: "$0.48";
|
|
115
|
+
}, {
|
|
116
|
+
readonly unit: "per M output tokens";
|
|
117
|
+
readonly price: "$0.030";
|
|
118
|
+
}];
|
|
119
|
+
};
|
|
90
120
|
}, {
|
|
91
121
|
readonly id: "ca54bcd6-0d98-4739-9b3b-5c8b4402193d";
|
|
92
122
|
readonly source: 1;
|
|
93
123
|
readonly name: "@cf/meta/llama-2-7b-chat-fp16";
|
|
94
124
|
readonly description: "Full precision (fp16) generative text model with 7 billion parameters from Meta";
|
|
125
|
+
readonly created_at: "2023-11-07 11:54:20.229";
|
|
95
126
|
readonly tags: readonly [];
|
|
96
127
|
readonly properties: {
|
|
97
128
|
readonly beta: true;
|
|
@@ -104,6 +135,7 @@ export declare const workersAiCatalog: {
|
|
|
104
135
|
readonly source: 1;
|
|
105
136
|
readonly name: "@cf/mistral/mistral-7b-instruct-v0.1";
|
|
106
137
|
readonly description: "Instruct fine-tuned version of the Mistral-7b generative text model with 7 billion parameters";
|
|
138
|
+
readonly created_at: "2023-11-07 11:54:20.229";
|
|
107
139
|
readonly tags: readonly [];
|
|
108
140
|
readonly properties: {
|
|
109
141
|
readonly beta: true;
|
|
@@ -116,6 +148,7 @@ export declare const workersAiCatalog: {
|
|
|
116
148
|
readonly source: 1;
|
|
117
149
|
readonly name: "@cf/mistral/mistral-7b-instruct-v0.2-lora";
|
|
118
150
|
readonly description: "The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2.";
|
|
151
|
+
readonly created_at: "2024-04-01 22:14:40.529";
|
|
119
152
|
readonly tags: readonly [];
|
|
120
153
|
readonly properties: {
|
|
121
154
|
readonly beta: true;
|
|
@@ -127,6 +160,7 @@ export declare const workersAiCatalog: {
|
|
|
127
160
|
readonly source: 1;
|
|
128
161
|
readonly name: "@cf/tinyllama/tinyllama-1.1b-chat-v1.0";
|
|
129
162
|
readonly description: "The TinyLlama project aims to pretrain a 1.1B Llama model on 3 trillion tokens. This is the chat model finetuned on top of TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T.";
|
|
163
|
+
readonly created_at: "2024-02-27 18:25:37.524";
|
|
130
164
|
readonly tags: readonly [];
|
|
131
165
|
readonly properties: {
|
|
132
166
|
readonly beta: true;
|
|
@@ -138,6 +172,7 @@ export declare const workersAiCatalog: {
|
|
|
138
172
|
readonly source: 2;
|
|
139
173
|
readonly name: "@hf/mistral/mistral-7b-instruct-v0.2";
|
|
140
174
|
readonly description: "The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2. Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1: 32k context window (vs 8k context in v0.1), rope-theta = 1e6, and no Sliding-Window Attention.";
|
|
175
|
+
readonly created_at: "2024-04-02 13:00:59.244";
|
|
141
176
|
readonly tags: readonly [];
|
|
142
177
|
readonly properties: {
|
|
143
178
|
readonly beta: true;
|
|
@@ -153,6 +188,7 @@ export declare const workersAiCatalog: {
|
|
|
153
188
|
readonly source: 1;
|
|
154
189
|
readonly name: "@cf/fblgit/una-cybertron-7b-v2-bf16";
|
|
155
190
|
readonly description: "Cybertron 7B v2 is a 7B MistralAI based model, best on it's series. It was trained with SFT, DPO and UNA (Unified Neural Alignment) on multiple datasets.";
|
|
191
|
+
readonly created_at: "2024-04-24 14:37:19.494";
|
|
156
192
|
readonly tags: readonly [];
|
|
157
193
|
readonly properties: {
|
|
158
194
|
readonly beta: true;
|
|
@@ -163,9 +199,16 @@ export declare const workersAiCatalog: {
|
|
|
163
199
|
readonly source: 1;
|
|
164
200
|
readonly name: "@cf/deepseek-ai/deepseek-r1-distill-qwen-32b";
|
|
165
201
|
readonly description: "DeepSeek-R1-Distill-Qwen-32B is a model distilled from DeepSeek-R1 based on Qwen2.5. It outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.";
|
|
202
|
+
readonly created_at: "2025-01-22 19:48:55.776";
|
|
166
203
|
readonly tags: readonly [];
|
|
167
204
|
readonly properties: {
|
|
168
|
-
readonly
|
|
205
|
+
readonly price: readonly [{
|
|
206
|
+
readonly unit: "per M input tokens";
|
|
207
|
+
readonly price: "$0.50";
|
|
208
|
+
}, {
|
|
209
|
+
readonly unit: "per M output tokens";
|
|
210
|
+
readonly price: "$4.88";
|
|
211
|
+
}];
|
|
169
212
|
readonly terms: "https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE";
|
|
170
213
|
};
|
|
171
214
|
}, {
|
|
@@ -173,6 +216,7 @@ export declare const workersAiCatalog: {
|
|
|
173
216
|
readonly source: 1;
|
|
174
217
|
readonly name: "@cf/thebloke/discolm-german-7b-v1-awq";
|
|
175
218
|
readonly description: "DiscoLM German 7b is a Mistral-based large language model with a focus on German-language applications. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.";
|
|
219
|
+
readonly created_at: "2024-02-27 18:23:05.178";
|
|
176
220
|
readonly tags: readonly [];
|
|
177
221
|
readonly properties: {
|
|
178
222
|
readonly beta: true;
|
|
@@ -184,6 +228,7 @@ export declare const workersAiCatalog: {
|
|
|
184
228
|
readonly source: 1;
|
|
185
229
|
readonly name: "@cf/meta/llama-2-7b-chat-int8";
|
|
186
230
|
readonly description: "Quantized (int8) generative text model with 7 billion parameters from Meta";
|
|
231
|
+
readonly created_at: "2023-09-25 19:21:11.898";
|
|
187
232
|
readonly tags: readonly [];
|
|
188
233
|
readonly properties: {
|
|
189
234
|
readonly context_window: 8192;
|
|
@@ -193,8 +238,16 @@ export declare const workersAiCatalog: {
|
|
|
193
238
|
readonly source: 1;
|
|
194
239
|
readonly name: "@cf/meta/llama-3.1-8b-instruct-fp8";
|
|
195
240
|
readonly description: "Llama 3.1 8B quantized to FP8 precision";
|
|
241
|
+
readonly created_at: "2024-07-25 17:28:43.328";
|
|
196
242
|
readonly tags: readonly [];
|
|
197
243
|
readonly properties: {
|
|
244
|
+
readonly price: readonly [{
|
|
245
|
+
readonly unit: "per M input tokens";
|
|
246
|
+
readonly price: "$0.15";
|
|
247
|
+
}, {
|
|
248
|
+
readonly unit: "per M output tokens";
|
|
249
|
+
readonly price: "$0.29";
|
|
250
|
+
}];
|
|
198
251
|
readonly context_window: 32000;
|
|
199
252
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
|
|
200
253
|
};
|
|
@@ -203,6 +256,7 @@ export declare const workersAiCatalog: {
|
|
|
203
256
|
readonly source: 2;
|
|
204
257
|
readonly name: "@hf/thebloke/mistral-7b-instruct-v0.1-awq";
|
|
205
258
|
readonly description: "Mistral 7B Instruct v0.1 AWQ is an efficient, accurate and blazing-fast low-bit weight quantized Mistral variant.";
|
|
259
|
+
readonly created_at: "2023-11-24 00:27:15.869";
|
|
206
260
|
readonly tags: readonly [];
|
|
207
261
|
readonly properties: {
|
|
208
262
|
readonly beta: true;
|
|
@@ -214,6 +268,7 @@ export declare const workersAiCatalog: {
|
|
|
214
268
|
readonly source: 1;
|
|
215
269
|
readonly name: "@cf/qwen/qwen1.5-7b-chat-awq";
|
|
216
270
|
readonly description: "Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.";
|
|
271
|
+
readonly created_at: "2024-02-27 18:24:11.709";
|
|
217
272
|
readonly tags: readonly [];
|
|
218
273
|
readonly properties: {
|
|
219
274
|
readonly beta: true;
|
|
@@ -225,8 +280,16 @@ export declare const workersAiCatalog: {
|
|
|
225
280
|
readonly source: 1;
|
|
226
281
|
readonly name: "@cf/meta/llama-3.2-1b-instruct";
|
|
227
282
|
readonly description: "The Llama 3.2 instruction-tuned text only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
|
|
283
|
+
readonly created_at: "2024-09-25 21:36:32.050";
|
|
228
284
|
readonly tags: readonly [];
|
|
229
285
|
readonly properties: {
|
|
286
|
+
readonly price: readonly [{
|
|
287
|
+
readonly unit: "per M input tokens";
|
|
288
|
+
readonly price: "$0.027";
|
|
289
|
+
}, {
|
|
290
|
+
readonly unit: "per M output tokens";
|
|
291
|
+
readonly price: "$0.20";
|
|
292
|
+
}];
|
|
230
293
|
readonly context_window: 60000;
|
|
231
294
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
|
|
232
295
|
};
|
|
@@ -235,6 +298,7 @@ export declare const workersAiCatalog: {
|
|
|
235
298
|
readonly source: 2;
|
|
236
299
|
readonly name: "@hf/thebloke/llama-2-13b-chat-awq";
|
|
237
300
|
readonly description: "Llama 2 13B Chat AWQ is an efficient, accurate and blazing-fast low-bit weight quantized Llama 2 variant.";
|
|
301
|
+
readonly created_at: "2023-11-24 00:27:15.869";
|
|
238
302
|
readonly tags: readonly [];
|
|
239
303
|
readonly properties: {
|
|
240
304
|
readonly beta: true;
|
|
@@ -246,6 +310,7 @@ export declare const workersAiCatalog: {
|
|
|
246
310
|
readonly source: 2;
|
|
247
311
|
readonly name: "@hf/thebloke/deepseek-coder-6.7b-base-awq";
|
|
248
312
|
readonly description: "Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese.";
|
|
313
|
+
readonly created_at: "2024-02-06 18:16:27.183";
|
|
249
314
|
readonly tags: readonly [];
|
|
250
315
|
readonly properties: {
|
|
251
316
|
readonly beta: true;
|
|
@@ -257,6 +322,7 @@ export declare const workersAiCatalog: {
|
|
|
257
322
|
readonly source: 1;
|
|
258
323
|
readonly name: "@cf/meta-llama/llama-2-7b-chat-hf-lora";
|
|
259
324
|
readonly description: "This is a Llama2 base model that Cloudflare dedicated for inference with LoRA adapters. Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 7B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format. ";
|
|
325
|
+
readonly created_at: "2024-04-02 00:17:18.579";
|
|
260
326
|
readonly tags: readonly [];
|
|
261
327
|
readonly properties: {
|
|
262
328
|
readonly beta: true;
|
|
@@ -268,8 +334,16 @@ export declare const workersAiCatalog: {
|
|
|
268
334
|
readonly source: 1;
|
|
269
335
|
readonly name: "@cf/meta/llama-3.3-70b-instruct-fp8-fast";
|
|
270
336
|
readonly description: "Llama 3.3 70B quantized to fp8 precision, optimized to be faster.";
|
|
337
|
+
readonly created_at: "2024-12-06 17:09:18.338";
|
|
271
338
|
readonly tags: readonly [];
|
|
272
339
|
readonly properties: {
|
|
340
|
+
readonly price: readonly [{
|
|
341
|
+
readonly unit: "per M input tokens";
|
|
342
|
+
readonly price: "$0.29";
|
|
343
|
+
}, {
|
|
344
|
+
readonly unit: "per M output tokens";
|
|
345
|
+
readonly price: "$2.25";
|
|
346
|
+
}];
|
|
273
347
|
readonly context_window: 24000;
|
|
274
348
|
readonly function_calling: true;
|
|
275
349
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_3/LICENSE";
|
|
@@ -279,6 +353,7 @@ export declare const workersAiCatalog: {
|
|
|
279
353
|
readonly source: 2;
|
|
280
354
|
readonly name: "@hf/thebloke/openhermes-2.5-mistral-7b-awq";
|
|
281
355
|
readonly description: "OpenHermes 2.5 Mistral 7B is a state of the art Mistral Fine-tune, a continuation of OpenHermes 2 model, which trained on additional code datasets.";
|
|
356
|
+
readonly created_at: "2024-02-06 18:04:22.846";
|
|
282
357
|
readonly tags: readonly [];
|
|
283
358
|
readonly properties: {
|
|
284
359
|
readonly beta: true;
|
|
@@ -289,6 +364,7 @@ export declare const workersAiCatalog: {
|
|
|
289
364
|
readonly source: 2;
|
|
290
365
|
readonly name: "@hf/thebloke/deepseek-coder-6.7b-instruct-awq";
|
|
291
366
|
readonly description: "Deepseek Coder is composed of a series of code language models, each trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese.";
|
|
367
|
+
readonly created_at: "2024-02-06 18:18:27.462";
|
|
292
368
|
readonly tags: readonly [];
|
|
293
369
|
readonly properties: {
|
|
294
370
|
readonly beta: true;
|
|
@@ -300,6 +376,7 @@ export declare const workersAiCatalog: {
|
|
|
300
376
|
readonly source: 1;
|
|
301
377
|
readonly name: "@cf/deepseek-ai/deepseek-math-7b-instruct";
|
|
302
378
|
readonly description: "DeepSeekMath-Instruct 7B is a mathematically instructed tuning model derived from DeepSeekMath-Base 7B. DeepSeekMath is initialized with DeepSeek-Coder-v1.5 7B and continues pre-training on math-related tokens sourced from Common Crawl, together with natural language and code data for 500B tokens.";
|
|
379
|
+
readonly created_at: "2024-02-27 17:54:17.459";
|
|
303
380
|
readonly tags: readonly [];
|
|
304
381
|
readonly properties: {
|
|
305
382
|
readonly beta: true;
|
|
@@ -312,6 +389,7 @@ export declare const workersAiCatalog: {
|
|
|
312
389
|
readonly source: 1;
|
|
313
390
|
readonly name: "@cf/tiiuae/falcon-7b-instruct";
|
|
314
391
|
readonly description: "Falcon-7B-Instruct is a 7B parameters causal decoder-only model built by TII based on Falcon-7B and finetuned on a mixture of chat/instruct datasets.";
|
|
392
|
+
readonly created_at: "2024-02-27 18:21:15.796";
|
|
315
393
|
readonly tags: readonly [];
|
|
316
394
|
readonly properties: {
|
|
317
395
|
readonly beta: true;
|
|
@@ -323,6 +401,7 @@ export declare const workersAiCatalog: {
|
|
|
323
401
|
readonly source: 2;
|
|
324
402
|
readonly name: "@hf/nousresearch/hermes-2-pro-mistral-7b";
|
|
325
403
|
readonly description: "Hermes 2 Pro on Mistral 7B is the new flagship 7B Hermes! Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.";
|
|
404
|
+
readonly created_at: "2024-04-01 23:45:53.800";
|
|
326
405
|
readonly tags: readonly [];
|
|
327
406
|
readonly properties: {
|
|
328
407
|
readonly beta: true;
|
|
@@ -335,8 +414,16 @@ export declare const workersAiCatalog: {
|
|
|
335
414
|
readonly source: 1;
|
|
336
415
|
readonly name: "@cf/meta/llama-3.1-8b-instruct";
|
|
337
416
|
readonly description: "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models. The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.";
|
|
417
|
+
readonly created_at: "2024-07-18 22:53:33.746";
|
|
338
418
|
readonly tags: readonly [];
|
|
339
419
|
readonly properties: {
|
|
420
|
+
readonly price: readonly [{
|
|
421
|
+
readonly unit: "per M input tokens";
|
|
422
|
+
readonly price: "$0.28";
|
|
423
|
+
}, {
|
|
424
|
+
readonly unit: "per M output tokens";
|
|
425
|
+
readonly price: "$0.83";
|
|
426
|
+
}];
|
|
340
427
|
readonly context_window: 7968;
|
|
341
428
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
|
|
342
429
|
};
|
|
@@ -345,8 +432,16 @@ export declare const workersAiCatalog: {
|
|
|
345
432
|
readonly source: 1;
|
|
346
433
|
readonly name: "@cf/meta/llama-3.1-8b-instruct-awq";
|
|
347
434
|
readonly description: "Quantized (int4) generative text model with 8 billion parameters from Meta.\n";
|
|
435
|
+
readonly created_at: "2024-07-25 17:46:04.304";
|
|
348
436
|
readonly tags: readonly [];
|
|
349
437
|
readonly properties: {
|
|
438
|
+
readonly price: readonly [{
|
|
439
|
+
readonly unit: "per M input tokens";
|
|
440
|
+
readonly price: "$0.12";
|
|
441
|
+
}, {
|
|
442
|
+
readonly unit: "per M output tokens";
|
|
443
|
+
readonly price: "$0.27";
|
|
444
|
+
}];
|
|
350
445
|
readonly context_window: 8192;
|
|
351
446
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_1/LICENSE";
|
|
352
447
|
};
|
|
@@ -355,6 +450,7 @@ export declare const workersAiCatalog: {
|
|
|
355
450
|
readonly source: 2;
|
|
356
451
|
readonly name: "@hf/thebloke/zephyr-7b-beta-awq";
|
|
357
452
|
readonly description: "Zephyr 7B Beta AWQ is an efficient, accurate and blazing-fast low-bit weight quantized Zephyr model variant.";
|
|
453
|
+
readonly created_at: "2023-11-24 00:27:15.869";
|
|
358
454
|
readonly tags: readonly [];
|
|
359
455
|
readonly properties: {
|
|
360
456
|
readonly beta: true;
|
|
@@ -366,6 +462,7 @@ export declare const workersAiCatalog: {
|
|
|
366
462
|
readonly source: 1;
|
|
367
463
|
readonly name: "@cf/google/gemma-7b-it-lora";
|
|
368
464
|
readonly description: " This is a Gemma-7B base model that Cloudflare dedicates for inference with LoRA adapters. Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.";
|
|
465
|
+
readonly created_at: "2024-04-02 00:20:19.633";
|
|
369
466
|
readonly tags: readonly [];
|
|
370
467
|
readonly properties: {
|
|
371
468
|
readonly beta: true;
|
|
@@ -377,6 +474,7 @@ export declare const workersAiCatalog: {
|
|
|
377
474
|
readonly source: 1;
|
|
378
475
|
readonly name: "@cf/qwen/qwen1.5-1.8b-chat";
|
|
379
476
|
readonly description: "Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud.";
|
|
477
|
+
readonly created_at: "2024-02-27 18:30:31.723";
|
|
380
478
|
readonly tags: readonly [];
|
|
381
479
|
readonly properties: {
|
|
382
480
|
readonly beta: true;
|
|
@@ -388,8 +486,16 @@ export declare const workersAiCatalog: {
|
|
|
388
486
|
readonly source: 1;
|
|
389
487
|
readonly name: "@cf/meta/llama-3-8b-instruct-awq";
|
|
390
488
|
readonly description: "Quantized (int4) generative text model with 8 billion parameters from Meta.";
|
|
489
|
+
readonly created_at: "2024-05-09 23:32:47.584";
|
|
391
490
|
readonly tags: readonly [];
|
|
392
491
|
readonly properties: {
|
|
492
|
+
readonly price: readonly [{
|
|
493
|
+
readonly unit: "per M input tokens";
|
|
494
|
+
readonly price: "$0.12";
|
|
495
|
+
}, {
|
|
496
|
+
readonly unit: "per M output tokens";
|
|
497
|
+
readonly price: "$0.27";
|
|
498
|
+
}];
|
|
393
499
|
readonly context_window: 8192;
|
|
394
500
|
readonly info: "https://llama.meta.com";
|
|
395
501
|
readonly terms: "https://llama.meta.com/llama3/license/#";
|
|
@@ -399,8 +505,16 @@ export declare const workersAiCatalog: {
|
|
|
399
505
|
readonly source: 1;
|
|
400
506
|
readonly name: "@cf/meta/llama-3.2-11b-vision-instruct";
|
|
401
507
|
readonly description: " The Llama 3.2-Vision instruction-tuned models are optimized for visual recognition, image reasoning, captioning, and answering general questions about an image.";
|
|
508
|
+
readonly created_at: "2024-09-25 05:36:04.547";
|
|
402
509
|
readonly tags: readonly [];
|
|
403
510
|
readonly properties: {
|
|
511
|
+
readonly price: readonly [{
|
|
512
|
+
readonly unit: "per M input tokens";
|
|
513
|
+
readonly price: "$0.049";
|
|
514
|
+
}, {
|
|
515
|
+
readonly unit: "per M output tokens";
|
|
516
|
+
readonly price: "$0.68";
|
|
517
|
+
}];
|
|
404
518
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/LICENSE";
|
|
405
519
|
};
|
|
406
520
|
}, {
|
|
@@ -408,6 +522,7 @@ export declare const workersAiCatalog: {
|
|
|
408
522
|
readonly source: 1;
|
|
409
523
|
readonly name: "@cf/defog/sqlcoder-7b-2";
|
|
410
524
|
readonly description: "This model is intended to be used by non-technical users to understand data inside their SQL databases. ";
|
|
525
|
+
readonly created_at: "2024-02-27 18:18:46.095";
|
|
411
526
|
readonly tags: readonly [];
|
|
412
527
|
readonly properties: {
|
|
413
528
|
readonly beta: true;
|
|
@@ -420,6 +535,7 @@ export declare const workersAiCatalog: {
|
|
|
420
535
|
readonly source: 1;
|
|
421
536
|
readonly name: "@cf/microsoft/phi-2";
|
|
422
537
|
readonly description: "Phi-2 is a Transformer-based model with a next-word prediction objective, trained on 1.4T tokens from multiple passes on a mixture of Synthetic and Web datasets for NLP and coding.";
|
|
538
|
+
readonly created_at: "2024-02-27 18:26:21.126";
|
|
423
539
|
readonly tags: readonly [];
|
|
424
540
|
readonly properties: {
|
|
425
541
|
readonly beta: true;
|
|
@@ -431,6 +547,7 @@ export declare const workersAiCatalog: {
|
|
|
431
547
|
readonly source: 2;
|
|
432
548
|
readonly name: "@hf/meta-llama/meta-llama-3-8b-instruct";
|
|
433
549
|
readonly description: "Generation over generation, Meta Llama 3 demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning.\t";
|
|
550
|
+
readonly created_at: "2024-05-22 18:21:04.371";
|
|
434
551
|
readonly tags: readonly [];
|
|
435
552
|
readonly properties: {
|
|
436
553
|
readonly context_window: 8192;
|
|
@@ -440,6 +557,7 @@ export declare const workersAiCatalog: {
|
|
|
440
557
|
readonly source: 2;
|
|
441
558
|
readonly name: "@hf/google/gemma-7b-it";
|
|
442
559
|
readonly description: "Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants.";
|
|
560
|
+
readonly created_at: "2024-04-01 23:51:35.866";
|
|
443
561
|
readonly tags: readonly [];
|
|
444
562
|
readonly properties: {
|
|
445
563
|
readonly beta: true;
|
|
@@ -453,6 +571,7 @@ export declare const workersAiCatalog: {
|
|
|
453
571
|
readonly source: 1;
|
|
454
572
|
readonly name: "@cf/qwen/qwen1.5-14b-chat-awq";
|
|
455
573
|
readonly description: "Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.";
|
|
574
|
+
readonly created_at: "2024-02-27 18:24:45.316";
|
|
456
575
|
readonly tags: readonly [];
|
|
457
576
|
readonly properties: {
|
|
458
577
|
readonly beta: true;
|
|
@@ -464,6 +583,7 @@ export declare const workersAiCatalog: {
|
|
|
464
583
|
readonly source: 1;
|
|
465
584
|
readonly name: "@cf/openchat/openchat-3.5-0106";
|
|
466
585
|
readonly description: "OpenChat is an innovative library of open-source language models, fine-tuned with C-RLFT - a strategy inspired by offline reinforcement learning.";
|
|
586
|
+
readonly created_at: "2024-02-27 18:20:39.169";
|
|
467
587
|
readonly tags: readonly [];
|
|
468
588
|
readonly properties: {
|
|
469
589
|
readonly beta: true;
|
|
@@ -480,6 +600,7 @@ export declare const workersAiCatalog: {
|
|
|
480
600
|
readonly source: 1;
|
|
481
601
|
readonly name: "@cf/baai/bge-m3";
|
|
482
602
|
readonly description: "Multi-Functionality, Multi-Linguality, and Multi-Granularity embeddings model.";
|
|
603
|
+
readonly created_at: "2024-05-22 19:27:09.781";
|
|
483
604
|
readonly tags: readonly [];
|
|
484
605
|
readonly properties: {};
|
|
485
606
|
}, {
|
|
@@ -487,6 +608,7 @@ export declare const workersAiCatalog: {
|
|
|
487
608
|
readonly source: 1;
|
|
488
609
|
readonly name: "@cf/baai/bge-small-en-v1.5";
|
|
489
610
|
readonly description: "BAAI general embedding (Small) model that transforms any given text into a 384-dimensional vector";
|
|
611
|
+
readonly created_at: "2023-11-07 15:43:58.042";
|
|
490
612
|
readonly tags: readonly [];
|
|
491
613
|
readonly properties: {
|
|
492
614
|
readonly beta: true;
|
|
@@ -499,6 +621,7 @@ export declare const workersAiCatalog: {
|
|
|
499
621
|
readonly source: 1;
|
|
500
622
|
readonly name: "@cf/baai/bge-base-en-v1.5";
|
|
501
623
|
readonly description: "BAAI general embedding (Base) model that transforms any given text into a 768-dimensional vector";
|
|
624
|
+
readonly created_at: "2023-09-25 19:21:11.898";
|
|
502
625
|
readonly tags: readonly [];
|
|
503
626
|
readonly properties: {
|
|
504
627
|
readonly beta: true;
|
|
@@ -511,6 +634,7 @@ export declare const workersAiCatalog: {
|
|
|
511
634
|
readonly source: 1;
|
|
512
635
|
readonly name: "@cf/baai/bge-large-en-v1.5";
|
|
513
636
|
readonly description: "BAAI general embedding (Large) model that transforms any given text into a 1024-dimensional vector";
|
|
637
|
+
readonly created_at: "2023-11-07 15:43:58.042";
|
|
514
638
|
readonly tags: readonly [];
|
|
515
639
|
readonly properties: {
|
|
516
640
|
readonly beta: true;
|
|
@@ -528,6 +652,7 @@ export declare const workersAiCatalog: {
|
|
|
528
652
|
readonly source: 1;
|
|
529
653
|
readonly name: "@cf/huggingface/distilbert-sst-2-int8";
|
|
530
654
|
readonly description: "Distilled BERT model that was finetuned on SST-2 for sentiment classification";
|
|
655
|
+
readonly created_at: "2023-09-25 19:21:11.898";
|
|
531
656
|
readonly tags: readonly [];
|
|
532
657
|
readonly properties: {
|
|
533
658
|
readonly beta: true;
|
|
@@ -538,6 +663,7 @@ export declare const workersAiCatalog: {
|
|
|
538
663
|
readonly source: 1;
|
|
539
664
|
readonly name: "@cf/baai/bge-reranker-base";
|
|
540
665
|
readonly description: "Different from embedding model, reranker uses question and document as input and directly output similarity instead of embedding. You can get a relevance score by inputting query and passage to the reranker. And the score can be mapped to a float value in [0,1] by sigmoid function.\n\n";
|
|
666
|
+
readonly created_at: "2025-02-14 12:28:19.009";
|
|
541
667
|
readonly tags: readonly [];
|
|
542
668
|
readonly properties: {};
|
|
543
669
|
}];
|
|
@@ -550,6 +676,7 @@ export declare const workersAiCatalog: {
|
|
|
550
676
|
readonly source: 1;
|
|
551
677
|
readonly name: "@cf/facebook/detr-resnet-50";
|
|
552
678
|
readonly description: "DEtection TRansformer (DETR) model trained end-to-end on COCO 2017 object detection (118k annotated images).";
|
|
679
|
+
readonly created_at: "2024-02-27 17:43:51.922";
|
|
553
680
|
readonly tags: readonly [];
|
|
554
681
|
readonly properties: {
|
|
555
682
|
readonly beta: true;
|
|
@@ -564,6 +691,7 @@ export declare const workersAiCatalog: {
|
|
|
564
691
|
readonly source: 1;
|
|
565
692
|
readonly name: "@cf/myshell-ai/melotts";
|
|
566
693
|
readonly description: "MeloTTS is a high-quality multi-lingual text-to-speech library by MyShell.ai.";
|
|
694
|
+
readonly created_at: "2024-07-19 15:51:04.819";
|
|
567
695
|
readonly tags: readonly [];
|
|
568
696
|
readonly properties: {};
|
|
569
697
|
}];
|
|
@@ -576,6 +704,7 @@ export declare const workersAiCatalog: {
|
|
|
576
704
|
readonly source: 1;
|
|
577
705
|
readonly name: "@cf/openai/whisper";
|
|
578
706
|
readonly description: "Whisper is a general-purpose speech recognition model. It is trained on a large dataset of diverse audio and is also a multitasking model that can perform multilingual speech recognition, speech translation, and language identification.";
|
|
707
|
+
readonly created_at: "2023-09-25 19:21:11.898";
|
|
579
708
|
readonly tags: readonly [];
|
|
580
709
|
readonly properties: {
|
|
581
710
|
readonly beta: true;
|
|
@@ -586,6 +715,7 @@ export declare const workersAiCatalog: {
|
|
|
586
715
|
readonly source: 1;
|
|
587
716
|
readonly name: "@cf/openai/whisper-tiny-en";
|
|
588
717
|
readonly description: "Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours of labelled data, Whisper models demonstrate a strong ability to generalize to many datasets and domains without the need for fine-tuning. This is the English-only version of the Whisper Tiny model which was trained on the task of speech recognition.";
|
|
718
|
+
readonly created_at: "2024-04-22 20:59:02.731";
|
|
589
719
|
readonly tags: readonly [];
|
|
590
720
|
readonly properties: {
|
|
591
721
|
readonly beta: true;
|
|
@@ -595,8 +725,14 @@ export declare const workersAiCatalog: {
|
|
|
595
725
|
readonly source: 1;
|
|
596
726
|
readonly name: "@cf/openai/whisper-large-v3-turbo";
|
|
597
727
|
readonly description: "Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. ";
|
|
728
|
+
readonly created_at: "2024-05-22 00:02:18.656";
|
|
598
729
|
readonly tags: readonly [];
|
|
599
|
-
readonly properties: {
|
|
730
|
+
readonly properties: {
|
|
731
|
+
readonly price: readonly [{
|
|
732
|
+
readonly unit: "per audio minute";
|
|
733
|
+
readonly price: "$0.00051";
|
|
734
|
+
}];
|
|
735
|
+
};
|
|
600
736
|
}];
|
|
601
737
|
};
|
|
602
738
|
readonly 'Image-to-Text': {
|
|
@@ -607,6 +743,7 @@ export declare const workersAiCatalog: {
|
|
|
607
743
|
readonly source: 1;
|
|
608
744
|
readonly name: "@cf/llava-hf/llava-1.5-7b-hf";
|
|
609
745
|
readonly description: "LLaVA is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-generated multimodal instruction-following data. It is an auto-regressive language model, based on the transformer architecture.";
|
|
746
|
+
readonly created_at: "2024-05-01 18:00:39.971";
|
|
610
747
|
readonly tags: readonly [];
|
|
611
748
|
readonly properties: {
|
|
612
749
|
readonly beta: true;
|
|
@@ -616,6 +753,7 @@ export declare const workersAiCatalog: {
|
|
|
616
753
|
readonly source: 1;
|
|
617
754
|
readonly name: "@cf/unum/uform-gen2-qwen-500m";
|
|
618
755
|
readonly description: "UForm-Gen is a small generative vision-language model primarily designed for Image Captioning and Visual Question Answering. The model was pre-trained on the internal image captioning dataset and fine-tuned on public instructions datasets: SVIT, LVIS, VQAs datasets.";
|
|
756
|
+
readonly created_at: "2024-02-27 18:28:52.485";
|
|
619
757
|
readonly tags: readonly [];
|
|
620
758
|
readonly properties: {
|
|
621
759
|
readonly beta: true;
|
|
@@ -631,6 +769,7 @@ export declare const workersAiCatalog: {
|
|
|
631
769
|
readonly source: 1;
|
|
632
770
|
readonly name: "@cf/runwayml/stable-diffusion-v1-5-inpainting";
|
|
633
771
|
readonly description: "Stable Diffusion Inpainting is a latent text-to-image diffusion model capable of generating photo-realistic images given any text input, with the extra capability of inpainting the pictures by using a mask.";
|
|
772
|
+
readonly created_at: "2024-02-27 17:23:57.528";
|
|
634
773
|
readonly tags: readonly [];
|
|
635
774
|
readonly properties: {
|
|
636
775
|
readonly beta: true;
|
|
@@ -642,6 +781,7 @@ export declare const workersAiCatalog: {
|
|
|
642
781
|
readonly source: 1;
|
|
643
782
|
readonly name: "@cf/black-forest-labs/flux-1-schnell";
|
|
644
783
|
readonly description: "FLUX.1 [schnell] is a 12 billion parameter rectified flow transformer capable of generating images from text descriptions. ";
|
|
784
|
+
readonly created_at: "2024-08-29 16:37:39.541";
|
|
645
785
|
readonly tags: readonly [];
|
|
646
786
|
readonly properties: {};
|
|
647
787
|
}, {
|
|
@@ -649,6 +789,7 @@ export declare const workersAiCatalog: {
|
|
|
649
789
|
readonly source: 1;
|
|
650
790
|
readonly name: "@cf/bytedance/stable-diffusion-xl-lightning";
|
|
651
791
|
readonly description: "SDXL-Lightning is a lightning-fast text-to-image generation model. It can generate high-quality 1024px images in a few steps.";
|
|
792
|
+
readonly created_at: "2024-02-27 17:41:29.578";
|
|
652
793
|
readonly tags: readonly [];
|
|
653
794
|
readonly properties: {
|
|
654
795
|
readonly beta: true;
|
|
@@ -659,6 +800,7 @@ export declare const workersAiCatalog: {
|
|
|
659
800
|
readonly source: 1;
|
|
660
801
|
readonly name: "@cf/lykon/dreamshaper-8-lcm";
|
|
661
802
|
readonly description: "Stable Diffusion model that has been fine-tuned to be better at photorealism without sacrificing range.";
|
|
803
|
+
readonly created_at: "2024-02-27 17:40:38.881";
|
|
662
804
|
readonly tags: readonly [];
|
|
663
805
|
readonly properties: {
|
|
664
806
|
readonly beta: true;
|
|
@@ -669,6 +811,7 @@ export declare const workersAiCatalog: {
|
|
|
669
811
|
readonly source: 1;
|
|
670
812
|
readonly name: "@cf/stabilityai/stable-diffusion-xl-base-1.0";
|
|
671
813
|
readonly description: "Diffusion-based text-to-image generative model by Stability AI. Generates and modify images based on text prompts.";
|
|
814
|
+
readonly created_at: "2023-11-10 10:54:43.694";
|
|
672
815
|
readonly tags: readonly [];
|
|
673
816
|
readonly properties: {
|
|
674
817
|
readonly beta: true;
|
|
@@ -680,6 +823,7 @@ export declare const workersAiCatalog: {
|
|
|
680
823
|
readonly source: 1;
|
|
681
824
|
readonly name: "@cf/runwayml/stable-diffusion-v1-5-img2img";
|
|
682
825
|
readonly description: "Stable Diffusion is a latent text-to-image diffusion model capable of generating photo-realistic images. Img2img generate a new image from an input image with Stable Diffusion. ";
|
|
826
|
+
readonly created_at: "2024-02-27 17:32:28.581";
|
|
683
827
|
readonly tags: readonly [];
|
|
684
828
|
readonly properties: {
|
|
685
829
|
readonly beta: true;
|
|
@@ -696,6 +840,7 @@ export declare const workersAiCatalog: {
|
|
|
696
840
|
readonly source: 1;
|
|
697
841
|
readonly name: "@cf/microsoft/resnet-50";
|
|
698
842
|
readonly description: "50 layers deep image classification CNN trained on more than 1M images from ImageNet";
|
|
843
|
+
readonly created_at: "2023-09-25 19:21:11.898";
|
|
699
844
|
readonly tags: readonly [];
|
|
700
845
|
readonly properties: {
|
|
701
846
|
readonly beta: true;
|
|
@@ -711,6 +856,7 @@ export declare const workersAiCatalog: {
|
|
|
711
856
|
readonly source: 1;
|
|
712
857
|
readonly name: "@cf/meta/m2m100-1.2b";
|
|
713
858
|
readonly description: "Multilingual encoder-decoder (seq-to-seq) model trained for Many-to-Many multilingual translation";
|
|
859
|
+
readonly created_at: "2023-09-25 19:21:11.898";
|
|
714
860
|
readonly tags: readonly [];
|
|
715
861
|
readonly properties: {
|
|
716
862
|
readonly beta: true;
|
|
@@ -728,6 +874,7 @@ export declare const workersAiCatalog: {
|
|
|
728
874
|
readonly source: 1;
|
|
729
875
|
readonly name: "@cf/facebook/bart-large-cnn";
|
|
730
876
|
readonly description: "BART is a transformer encoder-encoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. You can use this model for text summarization.";
|
|
877
|
+
readonly created_at: "2024-02-27 18:28:11.833";
|
|
731
878
|
readonly tags: readonly [];
|
|
732
879
|
readonly properties: {
|
|
733
880
|
readonly beta: true;
|