@chainfuse/types 2.10.16 → 2.10.18
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/ai-tools/azure/catalog.d.ts +0 -6
- package/dist/ai-tools/azure/catalog.js +0 -6
- package/dist/ai-tools/index.d.ts +39 -21
- package/dist/ai-tools/index.js +19 -9
- package/dist/ai-tools/workers-ai/catalog.d.ts +5 -4
- package/dist/ai-tools/workers-ai/catalog.js +5 -4
- package/dist/d0/index.d.ts +1 -0
- package/dist/d0/index.js +1 -1
- package/dist/d1/index.d.ts +1 -1
- package/package.json +4 -4
|
@@ -230,8 +230,6 @@ export declare const azureCatalog: readonly [{
|
|
|
230
230
|
readonly name: "gpt-4.1-nano";
|
|
231
231
|
}, {
|
|
232
232
|
readonly name: "gpt-4o";
|
|
233
|
-
readonly inputTokenCost: 0.00000275;
|
|
234
|
-
readonly outputTokenCost: 0.000011;
|
|
235
233
|
}, {
|
|
236
234
|
readonly name: "gpt-4o-mini";
|
|
237
235
|
readonly inputTokenCost: 1.65e-7;
|
|
@@ -634,8 +632,6 @@ export declare const azureCatalog: readonly [{
|
|
|
634
632
|
readonly name: "gpt-4.1-nano";
|
|
635
633
|
}, {
|
|
636
634
|
readonly name: "gpt-4o";
|
|
637
|
-
readonly inputTokenCost: 0.00000275;
|
|
638
|
-
readonly outputTokenCost: 0.000011;
|
|
639
635
|
}, {
|
|
640
636
|
readonly name: "gpt-4o-mini";
|
|
641
637
|
readonly inputTokenCost: 1.65e-7;
|
|
@@ -721,8 +717,6 @@ export declare const azureCatalog: readonly [{
|
|
|
721
717
|
readonly name: "gpt-4.1-nano";
|
|
722
718
|
}, {
|
|
723
719
|
readonly name: "gpt-4o";
|
|
724
|
-
readonly inputTokenCost: 0.00000275;
|
|
725
|
-
readonly outputTokenCost: 0.000011;
|
|
726
720
|
}, {
|
|
727
721
|
readonly name: "gpt-4o-mini";
|
|
728
722
|
readonly inputTokenCost: 1.65e-7;
|
|
@@ -300,8 +300,6 @@ export const azureCatalog = [
|
|
|
300
300
|
},
|
|
301
301
|
{
|
|
302
302
|
name: 'gpt-4o',
|
|
303
|
-
inputTokenCost: 0.00000275,
|
|
304
|
-
outputTokenCost: 0.000011,
|
|
305
303
|
},
|
|
306
304
|
{
|
|
307
305
|
name: 'gpt-4o-mini',
|
|
@@ -839,8 +837,6 @@ export const azureCatalog = [
|
|
|
839
837
|
},
|
|
840
838
|
{
|
|
841
839
|
name: 'gpt-4o',
|
|
842
|
-
inputTokenCost: 0.00000275,
|
|
843
|
-
outputTokenCost: 0.000011,
|
|
844
840
|
},
|
|
845
841
|
{
|
|
846
842
|
name: 'gpt-4o-mini',
|
|
@@ -956,8 +952,6 @@ export const azureCatalog = [
|
|
|
956
952
|
},
|
|
957
953
|
{
|
|
958
954
|
name: 'gpt-4o',
|
|
959
|
-
inputTokenCost: 0.00000275,
|
|
960
|
-
outputTokenCost: 0.000011,
|
|
961
955
|
},
|
|
962
956
|
{
|
|
963
957
|
name: 'gpt-4o-mini',
|
package/dist/ai-tools/index.d.ts
CHANGED
|
@@ -9,14 +9,20 @@ export interface Coordinate {
|
|
|
9
9
|
export declare namespace AiModels {
|
|
10
10
|
namespace LanguageModels {
|
|
11
11
|
enum Azure {
|
|
12
|
-
|
|
12
|
+
gpt41_nano = "azure:gpt-4.1-nano",
|
|
13
13
|
gpt4o_mini = "azure:gpt-4o-mini",
|
|
14
|
-
|
|
15
|
-
|
|
14
|
+
gpt41_mini = "azure:gpt-4.1-mini",
|
|
15
|
+
o1_mini = "azure:o1-mini",
|
|
16
|
+
o3_mini = "azure:o3-mini",
|
|
17
|
+
o4_mini = "azure:o4-mini",
|
|
18
|
+
gpt4o = "azure:gpt-4o",
|
|
19
|
+
gpt41 = "azure:gpt-4.1",
|
|
20
|
+
o1 = "azure:o1",
|
|
21
|
+
o3 = "azure:o3"
|
|
16
22
|
}
|
|
17
23
|
enum Anthropic {
|
|
18
|
-
|
|
19
|
-
|
|
24
|
+
sonnet = "anthropic:claude-4-sonnet-latest",
|
|
25
|
+
haiku = "anthropic:claude-4-opus-latest"
|
|
20
26
|
}
|
|
21
27
|
const Cloudflare: Readonly<CloudflareModelsEnum<"Text Generation">>;
|
|
22
28
|
const CloudflareFunctions: Readonly<CloudflareFunctionModelsEnum>;
|
|
@@ -32,18 +38,22 @@ export declare namespace AiModels {
|
|
|
32
38
|
gemini_pro_search = "google.generative-ai:gemini-2.5-pro-preview:search"
|
|
33
39
|
}
|
|
34
40
|
enum OpenAi {
|
|
35
|
-
|
|
41
|
+
gpt41_nano = "openai:gpt-4.1-nano",
|
|
36
42
|
gpt4o_mini = "openai:gpt-4o-mini",
|
|
43
|
+
gpt41_mini = "openai:gpt-4.1-mini",
|
|
37
44
|
o1_mini = "openai:o1-mini",
|
|
38
|
-
|
|
45
|
+
o3_mini = "openai:o3-mini",
|
|
46
|
+
o4_mini = "openai:o4-mini",
|
|
39
47
|
gpt4o = "openai:gpt-4o",
|
|
40
|
-
|
|
48
|
+
gpt41 = "openai:gpt-4.1",
|
|
49
|
+
o1 = "openai:o1",
|
|
50
|
+
o3 = "openai:o3"
|
|
41
51
|
}
|
|
42
52
|
}
|
|
43
53
|
namespace ImageModels {
|
|
44
54
|
enum Azure {
|
|
45
|
-
|
|
46
|
-
|
|
55
|
+
gpt1 = "azure:gpt-image-1",
|
|
56
|
+
dalle3 = "azure:dall-e-3"
|
|
47
57
|
}
|
|
48
58
|
const Cloudflare: Readonly<CloudflareModelsEnum<"Text-to-Image">>;
|
|
49
59
|
enum GoogleGenerativeAi {
|
|
@@ -51,8 +61,8 @@ export declare namespace AiModels {
|
|
|
51
61
|
imagen_fast = "google.generative-ai:imagen-3.0-fast-generate-001"
|
|
52
62
|
}
|
|
53
63
|
enum OpenAi {
|
|
54
|
-
|
|
55
|
-
|
|
64
|
+
openai = "azure:gpt-image-1",
|
|
65
|
+
dalle3 = "openai:dall-e-3"
|
|
56
66
|
}
|
|
57
67
|
}
|
|
58
68
|
namespace TextEmbeddingModels {
|
|
@@ -119,24 +129,34 @@ export declare const ZodLanguageModelValues: z4.ZodEnum<{
|
|
|
119
129
|
"workersai:@cf/meta/llama-4-scout-17b-16e-instruct": "workersai:@cf/meta/llama-4-scout-17b-16e-instruct";
|
|
120
130
|
"workersai:@cf/google/gemma-3-12b-it": "workersai:@cf/google/gemma-3-12b-it";
|
|
121
131
|
"workersai:@cf/qwen/qwq-32b": "workersai:@cf/qwen/qwq-32b";
|
|
122
|
-
"azure:gpt-
|
|
132
|
+
"azure:gpt-4.1-nano": AiModels.LanguageModels.Azure.gpt41_nano;
|
|
123
133
|
"azure:gpt-4o-mini": AiModels.LanguageModels.Azure.gpt4o_mini;
|
|
124
|
-
"azure:gpt-4-
|
|
134
|
+
"azure:gpt-4.1-mini": AiModels.LanguageModels.Azure.gpt41_mini;
|
|
135
|
+
"azure:o1-mini": AiModels.LanguageModels.Azure.o1_mini;
|
|
136
|
+
"azure:o3-mini": AiModels.LanguageModels.Azure.o3_mini;
|
|
137
|
+
"azure:o4-mini": AiModels.LanguageModels.Azure.o4_mini;
|
|
125
138
|
"azure:gpt-4o": AiModels.LanguageModels.Azure.gpt4o;
|
|
126
|
-
"
|
|
127
|
-
"
|
|
139
|
+
"azure:gpt-4.1": AiModels.LanguageModels.Azure.gpt41;
|
|
140
|
+
"azure:o1": AiModels.LanguageModels.Azure.o1;
|
|
141
|
+
"azure:o3": AiModels.LanguageModels.Azure.o3;
|
|
142
|
+
"anthropic:claude-4-sonnet-latest": AiModels.LanguageModels.Anthropic.sonnet;
|
|
143
|
+
"anthropic:claude-4-opus-latest": AiModels.LanguageModels.Anthropic.haiku;
|
|
128
144
|
"google.generative-ai:gemini-2.0-flash-lite": AiModels.LanguageModels.GoogleGenerativeAi.gemini_flash_lite;
|
|
129
145
|
"google.generative-ai:gemini-2.0-flash-lite:search": AiModels.LanguageModels.GoogleGenerativeAi.gemini_flash_lite_search;
|
|
130
146
|
"google.generative-ai:gemini-2.5-flash-preview": AiModels.LanguageModels.GoogleGenerativeAi.gemini_flash;
|
|
131
147
|
"google.generative-ai:gemini-2.5-flash-preview:search": AiModels.LanguageModels.GoogleGenerativeAi.gemini_flash_search;
|
|
132
148
|
"google.generative-ai:gemini-2.5-pro-preview": AiModels.LanguageModels.GoogleGenerativeAi.gemini_pro;
|
|
133
149
|
"google.generative-ai:gemini-2.5-pro-preview:search": AiModels.LanguageModels.GoogleGenerativeAi.gemini_pro_search;
|
|
134
|
-
"openai:gpt-
|
|
150
|
+
"openai:gpt-4.1-nano": AiModels.LanguageModels.OpenAi.gpt41_nano;
|
|
135
151
|
"openai:gpt-4o-mini": AiModels.LanguageModels.OpenAi.gpt4o_mini;
|
|
152
|
+
"openai:gpt-4.1-mini": AiModels.LanguageModels.OpenAi.gpt41_mini;
|
|
136
153
|
"openai:o1-mini": AiModels.LanguageModels.OpenAi.o1_mini;
|
|
137
|
-
"openai:
|
|
154
|
+
"openai:o3-mini": AiModels.LanguageModels.OpenAi.o3_mini;
|
|
155
|
+
"openai:o4-mini": AiModels.LanguageModels.OpenAi.o4_mini;
|
|
138
156
|
"openai:gpt-4o": AiModels.LanguageModels.OpenAi.gpt4o;
|
|
139
|
-
"openai:
|
|
157
|
+
"openai:gpt-4.1": AiModels.LanguageModels.OpenAi.gpt41;
|
|
158
|
+
"openai:o1": AiModels.LanguageModels.OpenAi.o1;
|
|
159
|
+
"openai:o3": AiModels.LanguageModels.OpenAi.o3;
|
|
140
160
|
}>;
|
|
141
161
|
export type LanguageModelValues = z4.infer<typeof ZodLanguageModelValues>;
|
|
142
162
|
export declare const ZodImageModelValues3: z3.ZodEnum<["workersai:@cf/runwayml/stable-diffusion-v1-5-inpainting" | "workersai:@cf/black-forest-labs/flux-1-schnell" | "workersai:@cf/bytedance/stable-diffusion-xl-lightning" | "workersai:@cf/lykon/dreamshaper-8-lcm" | "workersai:@cf/stabilityai/stable-diffusion-xl-base-1.0" | "workersai:@cf/runwayml/stable-diffusion-v1-5-img2img" | AiModels.ImageModels.Azure | AiModels.ImageModels.GoogleGenerativeAi | AiModels.ImageModels.OpenAi, ...("workersai:@cf/runwayml/stable-diffusion-v1-5-inpainting" | "workersai:@cf/black-forest-labs/flux-1-schnell" | "workersai:@cf/bytedance/stable-diffusion-xl-lightning" | "workersai:@cf/lykon/dreamshaper-8-lcm" | "workersai:@cf/stabilityai/stable-diffusion-xl-base-1.0" | "workersai:@cf/runwayml/stable-diffusion-v1-5-img2img" | AiModels.ImageModels.Azure | AiModels.ImageModels.GoogleGenerativeAi | AiModels.ImageModels.OpenAi)[]]>;
|
|
@@ -148,11 +168,9 @@ export declare const ZodImageModelValues: z4.ZodEnum<{
|
|
|
148
168
|
"workersai:@cf/stabilityai/stable-diffusion-xl-base-1.0": "workersai:@cf/stabilityai/stable-diffusion-xl-base-1.0";
|
|
149
169
|
"workersai:@cf/runwayml/stable-diffusion-v1-5-img2img": "workersai:@cf/runwayml/stable-diffusion-v1-5-img2img";
|
|
150
170
|
"azure:dall-e-3": AiModels.ImageModels.Azure.dalle3;
|
|
151
|
-
"azure:dall-e-2": AiModels.ImageModels.Azure.dalle2;
|
|
152
171
|
"google.generative-ai:imagen-3.0-generate-002": AiModels.ImageModels.GoogleGenerativeAi.imagen;
|
|
153
172
|
"google.generative-ai:imagen-3.0-fast-generate-001": AiModels.ImageModels.GoogleGenerativeAi.imagen_fast;
|
|
154
173
|
"openai:dall-e-3": AiModels.ImageModels.OpenAi.dalle3;
|
|
155
|
-
"openai:dall-e-2": AiModels.ImageModels.OpenAi.dalle2;
|
|
156
174
|
}>;
|
|
157
175
|
export type ImageModelValues = z4.infer<typeof ZodImageModelValues>;
|
|
158
176
|
export declare const ZodTextEmbeddingModelValues3: z3.ZodEnum<["workersai:@cf/baai/bge-m3" | "workersai:@cf/baai/bge-small-en-v1.5" | "workersai:@cf/baai/bge-base-en-v1.5" | "workersai:@cf/baai/bge-large-en-v1.5" | AiModels.TextEmbeddingModels.Azure | AiModels.TextEmbeddingModels.GoogleGenerativeAi.te4 | AiModels.TextEmbeddingModels.OpenAi, ...("workersai:@cf/baai/bge-m3" | "workersai:@cf/baai/bge-small-en-v1.5" | "workersai:@cf/baai/bge-base-en-v1.5" | "workersai:@cf/baai/bge-large-en-v1.5" | AiModels.TextEmbeddingModels.Azure | AiModels.TextEmbeddingModels.GoogleGenerativeAi.te4 | AiModels.TextEmbeddingModels.OpenAi)[]]>;
|
package/dist/ai-tools/index.js
CHANGED
|
@@ -7,15 +7,21 @@ export var AiModels;
|
|
|
7
7
|
(function (LanguageModels) {
|
|
8
8
|
let Azure;
|
|
9
9
|
(function (Azure) {
|
|
10
|
-
Azure["
|
|
10
|
+
Azure["gpt41_nano"] = "azure:gpt-4.1-nano";
|
|
11
11
|
Azure["gpt4o_mini"] = "azure:gpt-4o-mini";
|
|
12
|
-
Azure["
|
|
12
|
+
Azure["gpt41_mini"] = "azure:gpt-4.1-mini";
|
|
13
|
+
Azure["o1_mini"] = "azure:o1-mini";
|
|
14
|
+
Azure["o3_mini"] = "azure:o3-mini";
|
|
15
|
+
Azure["o4_mini"] = "azure:o4-mini";
|
|
13
16
|
Azure["gpt4o"] = "azure:gpt-4o";
|
|
17
|
+
Azure["gpt41"] = "azure:gpt-4.1";
|
|
18
|
+
Azure["o1"] = "azure:o1";
|
|
19
|
+
Azure["o3"] = "azure:o3";
|
|
14
20
|
})(Azure = LanguageModels.Azure || (LanguageModels.Azure = {}));
|
|
15
21
|
let Anthropic;
|
|
16
22
|
(function (Anthropic) {
|
|
17
|
-
Anthropic["
|
|
18
|
-
Anthropic["
|
|
23
|
+
Anthropic["sonnet"] = "anthropic:claude-4-sonnet-latest";
|
|
24
|
+
Anthropic["haiku"] = "anthropic:claude-4-opus-latest";
|
|
19
25
|
})(Anthropic = LanguageModels.Anthropic || (LanguageModels.Anthropic = {}));
|
|
20
26
|
// export const CloudflareSummary = Object.freeze(Object.fromEntries(enabledCloudflareLlmSummaryProviders.map((model) => [model, `workersai:${model}`])) as unknown as CloudflareModelsEnum<'Summarization'>);
|
|
21
27
|
// export const CloudflareClassification = Object.freeze(Object.fromEntries(enabledCloudflareLlmClassificationProviders.map((model) => [model, `workersai:${model}`])) as unknown as CloudflareModelsEnum<'Text Classification'>);
|
|
@@ -35,20 +41,24 @@ export var AiModels;
|
|
|
35
41
|
})(GoogleGenerativeAi = LanguageModels.GoogleGenerativeAi || (LanguageModels.GoogleGenerativeAi = {}));
|
|
36
42
|
let OpenAi;
|
|
37
43
|
(function (OpenAi) {
|
|
38
|
-
OpenAi["
|
|
44
|
+
OpenAi["gpt41_nano"] = "openai:gpt-4.1-nano";
|
|
39
45
|
OpenAi["gpt4o_mini"] = "openai:gpt-4o-mini";
|
|
46
|
+
OpenAi["gpt41_mini"] = "openai:gpt-4.1-mini";
|
|
40
47
|
OpenAi["o1_mini"] = "openai:o1-mini";
|
|
41
|
-
OpenAi["
|
|
48
|
+
OpenAi["o3_mini"] = "openai:o3-mini";
|
|
49
|
+
OpenAi["o4_mini"] = "openai:o4-mini";
|
|
42
50
|
OpenAi["gpt4o"] = "openai:gpt-4o";
|
|
43
|
-
OpenAi["
|
|
51
|
+
OpenAi["gpt41"] = "openai:gpt-4.1";
|
|
52
|
+
OpenAi["o1"] = "openai:o1";
|
|
53
|
+
OpenAi["o3"] = "openai:o3";
|
|
44
54
|
})(OpenAi = LanguageModels.OpenAi || (LanguageModels.OpenAi = {}));
|
|
45
55
|
})(LanguageModels = AiModels.LanguageModels || (AiModels.LanguageModels = {}));
|
|
46
56
|
let ImageModels;
|
|
47
57
|
(function (ImageModels) {
|
|
48
58
|
let Azure;
|
|
49
59
|
(function (Azure) {
|
|
60
|
+
Azure["gpt1"] = "azure:gpt-image-1";
|
|
50
61
|
Azure["dalle3"] = "azure:dall-e-3";
|
|
51
|
-
Azure["dalle2"] = "azure:dall-e-2";
|
|
52
62
|
})(Azure = ImageModels.Azure || (ImageModels.Azure = {}));
|
|
53
63
|
ImageModels.Cloudflare = Object.freeze(Object.fromEntries(enabledCloudflareLlmImageProviders.map((model) => [model, `workersai:${model}`])));
|
|
54
64
|
let GoogleGenerativeAi;
|
|
@@ -58,8 +68,8 @@ export var AiModels;
|
|
|
58
68
|
})(GoogleGenerativeAi = ImageModels.GoogleGenerativeAi || (ImageModels.GoogleGenerativeAi = {}));
|
|
59
69
|
let OpenAi;
|
|
60
70
|
(function (OpenAi) {
|
|
71
|
+
OpenAi["openai"] = "azure:gpt-image-1";
|
|
61
72
|
OpenAi["dalle3"] = "openai:dall-e-3";
|
|
62
|
-
OpenAi["dalle2"] = "openai:dall-e-2";
|
|
63
73
|
})(OpenAi = ImageModels.OpenAi || (ImageModels.OpenAi = {}));
|
|
64
74
|
})(ImageModels = AiModels.ImageModels || (AiModels.ImageModels = {}));
|
|
65
75
|
let TextEmbeddingModels;
|
|
@@ -567,6 +567,7 @@ export declare const workersAiCatalog: {
|
|
|
567
567
|
readonly created_at: "2024-09-25 05:36:04.547";
|
|
568
568
|
readonly tags: readonly [];
|
|
569
569
|
readonly properties: {
|
|
570
|
+
readonly context_window: 128000;
|
|
570
571
|
readonly price: readonly [{
|
|
571
572
|
readonly unit: "per M input tokens";
|
|
572
573
|
readonly price: 0.049;
|
|
@@ -660,6 +661,7 @@ export declare const workersAiCatalog: {
|
|
|
660
661
|
readonly created_at: "2025-04-05 20:25:56.137";
|
|
661
662
|
readonly tags: readonly [];
|
|
662
663
|
readonly properties: {
|
|
664
|
+
readonly context_window: 131000;
|
|
663
665
|
readonly price: readonly [{
|
|
664
666
|
readonly unit: "per M input tokens";
|
|
665
667
|
readonly price: 0.27;
|
|
@@ -669,7 +671,6 @@ export declare const workersAiCatalog: {
|
|
|
669
671
|
readonly price: 0.85;
|
|
670
672
|
readonly currency: "USD";
|
|
671
673
|
}];
|
|
672
|
-
readonly context_window: 131000;
|
|
673
674
|
readonly function_calling: true;
|
|
674
675
|
readonly terms: "https://github.com/meta-llama/llama-models/blob/main/models/llama4/LICENSE";
|
|
675
676
|
};
|
|
@@ -827,11 +828,11 @@ export declare const workersAiCatalog: {
|
|
|
827
828
|
readonly id: "9c178979-90d9-49d8-9e2c-0f1cf01815d4";
|
|
828
829
|
readonly description: "Object detection models can detect instances of objects like persons, faces, license plates, or others in an image. This task takes an image as input and returns a list of detected objects, each one containing a label, a probability score, and its surrounding box coordinates.";
|
|
829
830
|
readonly models: readonly [{
|
|
830
|
-
readonly id: "
|
|
831
|
+
readonly id: "d71b454b-1919-4cc7-a615-8827f34f72b3";
|
|
831
832
|
readonly source: 1;
|
|
832
|
-
readonly name: "@cf/facebook/detr-resnet-50";
|
|
833
|
+
readonly name: "@cf/facebook/triton-detr-resnet-50";
|
|
833
834
|
readonly description: "DEtection TRansformer (DETR) model trained end-to-end on COCO 2017 object detection (118k annotated images).";
|
|
834
|
-
readonly created_at: "
|
|
835
|
+
readonly created_at: "2025-05-29 11:20:49.014";
|
|
835
836
|
readonly tags: readonly [];
|
|
836
837
|
readonly properties: {
|
|
837
838
|
readonly beta: true;
|
|
@@ -644,6 +644,7 @@ export const workersAiCatalog = {
|
|
|
644
644
|
created_at: '2024-09-25 05:36:04.547',
|
|
645
645
|
tags: [],
|
|
646
646
|
properties: {
|
|
647
|
+
context_window: 128000,
|
|
647
648
|
price: [
|
|
648
649
|
{
|
|
649
650
|
unit: 'per M input tokens',
|
|
@@ -747,6 +748,7 @@ export const workersAiCatalog = {
|
|
|
747
748
|
created_at: '2025-04-05 20:25:56.137',
|
|
748
749
|
tags: [],
|
|
749
750
|
properties: {
|
|
751
|
+
context_window: 131000,
|
|
750
752
|
price: [
|
|
751
753
|
{
|
|
752
754
|
unit: 'per M input tokens',
|
|
@@ -759,7 +761,6 @@ export const workersAiCatalog = {
|
|
|
759
761
|
currency: 'USD',
|
|
760
762
|
},
|
|
761
763
|
],
|
|
762
|
-
context_window: 131000,
|
|
763
764
|
function_calling: true,
|
|
764
765
|
terms: 'https://github.com/meta-llama/llama-models/blob/main/models/llama4/LICENSE',
|
|
765
766
|
},
|
|
@@ -947,11 +948,11 @@ export const workersAiCatalog = {
|
|
|
947
948
|
description: 'Object detection models can detect instances of objects like persons, faces, license plates, or others in an image. This task takes an image as input and returns a list of detected objects, each one containing a label, a probability score, and its surrounding box coordinates.',
|
|
948
949
|
models: [
|
|
949
950
|
{
|
|
950
|
-
id: '
|
|
951
|
+
id: 'd71b454b-1919-4cc7-a615-8827f34f72b3',
|
|
951
952
|
source: 1,
|
|
952
|
-
name: '@cf/facebook/detr-resnet-50',
|
|
953
|
+
name: '@cf/facebook/triton-detr-resnet-50',
|
|
953
954
|
description: 'DEtection TRansformer (DETR) model trained end-to-end on COCO 2017 object detection (118k annotated images).',
|
|
954
|
-
created_at: '
|
|
955
|
+
created_at: '2025-05-29 11:20:49.014',
|
|
955
956
|
tags: [],
|
|
956
957
|
properties: {
|
|
957
958
|
beta: true,
|
package/dist/d0/index.d.ts
CHANGED
package/dist/d0/index.js
CHANGED
|
@@ -8,7 +8,7 @@ const PrefixedUuidRaw3 = await import('zod/v3').then(({ z }) => z
|
|
|
8
8
|
.min(38)
|
|
9
9
|
.max(40)
|
|
10
10
|
.regex(new RegExp(/^((d|t|u)_)?[0-9a-f]{8}-[0-9a-f]{4}-7[0-9a-f]{3}-[0-9a-f]{4}-[0-9a-f]{12}(_p)?$/i)));
|
|
11
|
-
const PrefixedUuidRaw = await import('zod/v4').then(({ z }) => z
|
|
11
|
+
export const PrefixedUuidRaw = await import('zod/v4').then(({ z }) => z
|
|
12
12
|
.string()
|
|
13
13
|
.trim()
|
|
14
14
|
.toLowerCase()
|
package/dist/d1/index.d.ts
CHANGED
|
@@ -3,7 +3,7 @@ export type PrefixedUuid = `${'t_' | 'd_' | 'u_'}${UuidExport['utf8']}${'' | '_p
|
|
|
3
3
|
export interface UuidExport {
|
|
4
4
|
utf8: ReturnType<typeof crypto.randomUUID>;
|
|
5
5
|
hex: string;
|
|
6
|
-
blob: Buffer
|
|
6
|
+
blob: Buffer['buffer'] | (typeof Uint8Array)['prototype']['buffer'];
|
|
7
7
|
base64: string;
|
|
8
8
|
base64url: string;
|
|
9
9
|
}
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@chainfuse/types",
|
|
3
|
-
"version": "2.10.
|
|
3
|
+
"version": "2.10.18",
|
|
4
4
|
"description": "",
|
|
5
5
|
"author": "ChainFuse",
|
|
6
6
|
"homepage": "https://github.com/ChainFuse/packages/tree/main/packages/types#readme",
|
|
@@ -99,11 +99,11 @@
|
|
|
99
99
|
"prettier": "@demosjarco/prettier-config",
|
|
100
100
|
"dependencies": {
|
|
101
101
|
"validator": "^13.15.15",
|
|
102
|
-
"zod": "^3.25.
|
|
102
|
+
"zod": "^3.25.42"
|
|
103
103
|
},
|
|
104
104
|
"devDependencies": {
|
|
105
|
-
"@cloudflare/workers-types": "^4.
|
|
105
|
+
"@cloudflare/workers-types": "^4.20250529.0",
|
|
106
106
|
"@types/validator": "^13.15.1"
|
|
107
107
|
},
|
|
108
|
-
"gitHead": "
|
|
108
|
+
"gitHead": "e7723806e2f740dd24652028891ddc4aa4fa0e57"
|
|
109
109
|
}
|