@cloudflare/workers-types 4.20250402.0 → 4.20250403.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/2021-11-03/index.d.ts +128 -3
- package/2021-11-03/index.ts +130 -3
- package/2022-01-31/index.d.ts +128 -3
- package/2022-01-31/index.ts +130 -3
- package/2022-03-21/index.d.ts +128 -3
- package/2022-03-21/index.ts +130 -3
- package/2022-08-04/index.d.ts +128 -3
- package/2022-08-04/index.ts +130 -3
- package/2022-10-31/index.d.ts +128 -3
- package/2022-10-31/index.ts +130 -3
- package/2022-11-30/index.d.ts +128 -3
- package/2022-11-30/index.ts +130 -3
- package/2023-03-01/index.d.ts +128 -3
- package/2023-03-01/index.ts +130 -3
- package/2023-07-01/index.d.ts +128 -3
- package/2023-07-01/index.ts +130 -3
- package/experimental/index.d.ts +129 -3
- package/experimental/index.ts +131 -3
- package/index.d.ts +128 -3
- package/index.ts +130 -3
- package/oldest/index.d.ts +128 -3
- package/oldest/index.ts +130 -3
- package/package.json +1 -1
package/2021-11-03/index.d.ts
CHANGED
|
@@ -3491,6 +3491,28 @@ declare abstract class BaseAiImageToText {
|
|
|
3491
3491
|
inputs: AiImageToTextInput;
|
|
3492
3492
|
postProcessedOutputs: AiImageToTextOutput;
|
|
3493
3493
|
}
|
|
3494
|
+
type AiImageTextToTextInput = {
|
|
3495
|
+
image: string;
|
|
3496
|
+
prompt?: string;
|
|
3497
|
+
max_tokens?: number;
|
|
3498
|
+
temperature?: number;
|
|
3499
|
+
ignore_eos?: boolean;
|
|
3500
|
+
top_p?: number;
|
|
3501
|
+
top_k?: number;
|
|
3502
|
+
seed?: number;
|
|
3503
|
+
repetition_penalty?: number;
|
|
3504
|
+
frequency_penalty?: number;
|
|
3505
|
+
presence_penalty?: number;
|
|
3506
|
+
raw?: boolean;
|
|
3507
|
+
messages?: RoleScopedChatInput[];
|
|
3508
|
+
};
|
|
3509
|
+
type AiImageTextToTextOutput = {
|
|
3510
|
+
description: string;
|
|
3511
|
+
};
|
|
3512
|
+
declare abstract class BaseAiImageTextToText {
|
|
3513
|
+
inputs: AiImageTextToTextInput;
|
|
3514
|
+
postProcessedOutputs: AiImageTextToTextOutput;
|
|
3515
|
+
}
|
|
3494
3516
|
type AiObjectDetectionInput = {
|
|
3495
3517
|
image: number[];
|
|
3496
3518
|
};
|
|
@@ -3901,6 +3923,72 @@ declare abstract class Base_Ai_Cf_Openai_Whisper_Large_V3_Turbo {
|
|
|
3901
3923
|
inputs: Ai_Cf_Openai_Whisper_Large_V3_Turbo_Input;
|
|
3902
3924
|
postProcessedOutputs: Ai_Cf_Openai_Whisper_Large_V3_Turbo_Output;
|
|
3903
3925
|
}
|
|
3926
|
+
type Ai_Cf_Baai_Bge_M3_Input = BGEM3InputQueryAndContexts | BGEM3InputEmbedding;
|
|
3927
|
+
interface BGEM3InputQueryAndContexts {
|
|
3928
|
+
/**
|
|
3929
|
+
* A query you wish to perform against the provided contexts. If no query is provided the model with respond with embeddings for contexts
|
|
3930
|
+
*/
|
|
3931
|
+
query?: string;
|
|
3932
|
+
/**
|
|
3933
|
+
* List of provided contexts. Note that the index in this array is important, as the response will refer to it.
|
|
3934
|
+
*/
|
|
3935
|
+
contexts: {
|
|
3936
|
+
/**
|
|
3937
|
+
* One of the provided context content
|
|
3938
|
+
*/
|
|
3939
|
+
text?: string;
|
|
3940
|
+
}[];
|
|
3941
|
+
/**
|
|
3942
|
+
* When provided with too long context should the model error out or truncate the context to fit?
|
|
3943
|
+
*/
|
|
3944
|
+
truncate_inputs?: boolean;
|
|
3945
|
+
}
|
|
3946
|
+
interface BGEM3InputEmbedding {
|
|
3947
|
+
text: string | string[];
|
|
3948
|
+
/**
|
|
3949
|
+
* When provided with too long context should the model error out or truncate the context to fit?
|
|
3950
|
+
*/
|
|
3951
|
+
truncate_inputs?: boolean;
|
|
3952
|
+
}
|
|
3953
|
+
type Ai_Cf_Baai_Bge_M3_Output =
|
|
3954
|
+
| BGEM3OuputQuery
|
|
3955
|
+
| BGEM3OutputEmbeddingForContexts
|
|
3956
|
+
| BGEM3OuputEmbedding;
|
|
3957
|
+
interface BGEM3OuputQuery {
|
|
3958
|
+
response?: {
|
|
3959
|
+
/**
|
|
3960
|
+
* Index of the context in the request
|
|
3961
|
+
*/
|
|
3962
|
+
id?: number;
|
|
3963
|
+
/**
|
|
3964
|
+
* Score of the context under the index.
|
|
3965
|
+
*/
|
|
3966
|
+
score?: number;
|
|
3967
|
+
}[];
|
|
3968
|
+
}
|
|
3969
|
+
interface BGEM3OutputEmbeddingForContexts {
|
|
3970
|
+
response?: number[][];
|
|
3971
|
+
shape?: number[];
|
|
3972
|
+
/**
|
|
3973
|
+
* The pooling method used in the embedding process.
|
|
3974
|
+
*/
|
|
3975
|
+
pooling?: "mean" | "cls";
|
|
3976
|
+
}
|
|
3977
|
+
interface BGEM3OuputEmbedding {
|
|
3978
|
+
shape?: number[];
|
|
3979
|
+
/**
|
|
3980
|
+
* Embeddings of the requested text values
|
|
3981
|
+
*/
|
|
3982
|
+
data?: number[][];
|
|
3983
|
+
/**
|
|
3984
|
+
* The pooling method used in the embedding process.
|
|
3985
|
+
*/
|
|
3986
|
+
pooling?: "mean" | "cls";
|
|
3987
|
+
}
|
|
3988
|
+
declare abstract class Base_Ai_Cf_Baai_Bge_M3 {
|
|
3989
|
+
inputs: Ai_Cf_Baai_Bge_M3_Input;
|
|
3990
|
+
postProcessedOutputs: Ai_Cf_Baai_Bge_M3_Output;
|
|
3991
|
+
}
|
|
3904
3992
|
interface Ai_Cf_Black_Forest_Labs_Flux_1_Schnell_Input {
|
|
3905
3993
|
/**
|
|
3906
3994
|
* A text description of the image you want to generate.
|
|
@@ -4211,6 +4299,40 @@ declare abstract class Base_Ai_Cf_Meta_Llama_Guard_3_8B {
|
|
|
4211
4299
|
inputs: Ai_Cf_Meta_Llama_Guard_3_8B_Input;
|
|
4212
4300
|
postProcessedOutputs: Ai_Cf_Meta_Llama_Guard_3_8B_Output;
|
|
4213
4301
|
}
|
|
4302
|
+
interface Ai_Cf_Baai_Bge_Reranker_Base_Input {
|
|
4303
|
+
/**
|
|
4304
|
+
* A query you wish to perform against the provided contexts.
|
|
4305
|
+
*/
|
|
4306
|
+
/**
|
|
4307
|
+
* Number of returned results starting with the best score.
|
|
4308
|
+
*/
|
|
4309
|
+
top_k?: number;
|
|
4310
|
+
/**
|
|
4311
|
+
* List of provided contexts. Note that the index in this array is important, as the response will refer to it.
|
|
4312
|
+
*/
|
|
4313
|
+
contexts: {
|
|
4314
|
+
/**
|
|
4315
|
+
* One of the provided context content
|
|
4316
|
+
*/
|
|
4317
|
+
text?: string;
|
|
4318
|
+
}[];
|
|
4319
|
+
}
|
|
4320
|
+
interface Ai_Cf_Baai_Bge_Reranker_Base_Output {
|
|
4321
|
+
response?: {
|
|
4322
|
+
/**
|
|
4323
|
+
* Index of the context in the request
|
|
4324
|
+
*/
|
|
4325
|
+
id?: number;
|
|
4326
|
+
/**
|
|
4327
|
+
* Score of the context under the index.
|
|
4328
|
+
*/
|
|
4329
|
+
score?: number;
|
|
4330
|
+
}[];
|
|
4331
|
+
}
|
|
4332
|
+
declare abstract class Base_Ai_Cf_Baai_Bge_Reranker_Base {
|
|
4333
|
+
inputs: Ai_Cf_Baai_Bge_Reranker_Base_Input;
|
|
4334
|
+
postProcessedOutputs: Ai_Cf_Baai_Bge_Reranker_Base_Output;
|
|
4335
|
+
}
|
|
4214
4336
|
interface AiModels {
|
|
4215
4337
|
"@cf/huggingface/distilbert-sst-2-int8": BaseAiTextClassification;
|
|
4216
4338
|
"@cf/stabilityai/stable-diffusion-xl-base-1.0": BaseAiTextToImage;
|
|
@@ -4218,6 +4340,7 @@ interface AiModels {
|
|
|
4218
4340
|
"@cf/runwayml/stable-diffusion-v1-5-img2img": BaseAiTextToImage;
|
|
4219
4341
|
"@cf/lykon/dreamshaper-8-lcm": BaseAiTextToImage;
|
|
4220
4342
|
"@cf/bytedance/stable-diffusion-xl-lightning": BaseAiTextToImage;
|
|
4343
|
+
"@cf/myshell-ai/melotts": BaseAiTextToSpeech;
|
|
4221
4344
|
"@cf/baai/bge-base-en-v1.5": BaseAiTextEmbeddings;
|
|
4222
4345
|
"@cf/baai/bge-small-en-v1.5": BaseAiTextEmbeddings;
|
|
4223
4346
|
"@cf/baai/bge-large-en-v1.5": BaseAiTextEmbeddings;
|
|
@@ -4271,9 +4394,11 @@ interface AiModels {
|
|
|
4271
4394
|
"@cf/unum/uform-gen2-qwen-500m": Base_Ai_Cf_Unum_Uform_Gen2_Qwen_500M;
|
|
4272
4395
|
"@cf/openai/whisper-tiny-en": Base_Ai_Cf_Openai_Whisper_Tiny_En;
|
|
4273
4396
|
"@cf/openai/whisper-large-v3-turbo": Base_Ai_Cf_Openai_Whisper_Large_V3_Turbo;
|
|
4397
|
+
"@cf/baai/bge-m3": Base_Ai_Cf_Baai_Bge_M3;
|
|
4274
4398
|
"@cf/black-forest-labs/flux-1-schnell": Base_Ai_Cf_Black_Forest_Labs_Flux_1_Schnell;
|
|
4275
4399
|
"@cf/meta/llama-3.2-11b-vision-instruct": Base_Ai_Cf_Meta_Llama_3_2_11B_Vision_Instruct;
|
|
4276
4400
|
"@cf/meta/llama-guard-3-8b": Base_Ai_Cf_Meta_Llama_Guard_3_8B;
|
|
4401
|
+
"@cf/baai/bge-reranker-base": Base_Ai_Cf_Baai_Bge_Reranker_Base;
|
|
4277
4402
|
}
|
|
4278
4403
|
type AiOptions = {
|
|
4279
4404
|
gateway?: GatewayOptions;
|
|
@@ -4331,8 +4456,8 @@ declare abstract class Ai<AiModelList extends AiModelListType = AiModels> {
|
|
|
4331
4456
|
? Response
|
|
4332
4457
|
: AiModelList[Name]["postProcessedOutputs"]
|
|
4333
4458
|
>;
|
|
4334
|
-
|
|
4335
|
-
|
|
4459
|
+
models(params?: AiModelsSearchParams): Promise<AiModelsSearchObject[]>;
|
|
4460
|
+
toMarkdown(
|
|
4336
4461
|
files: {
|
|
4337
4462
|
name: string;
|
|
4338
4463
|
blob: Blob;
|
|
@@ -4342,7 +4467,7 @@ declare abstract class Ai<AiModelList extends AiModelListType = AiModels> {
|
|
|
4342
4467
|
extraHeaders?: object;
|
|
4343
4468
|
},
|
|
4344
4469
|
): Promise<ConversionResponse[]>;
|
|
4345
|
-
|
|
4470
|
+
toMarkdown(
|
|
4346
4471
|
files: {
|
|
4347
4472
|
name: string;
|
|
4348
4473
|
blob: Blob;
|
package/2021-11-03/index.ts
CHANGED
|
@@ -3503,6 +3503,28 @@ export declare abstract class BaseAiImageToText {
|
|
|
3503
3503
|
inputs: AiImageToTextInput;
|
|
3504
3504
|
postProcessedOutputs: AiImageToTextOutput;
|
|
3505
3505
|
}
|
|
3506
|
+
export type AiImageTextToTextInput = {
|
|
3507
|
+
image: string;
|
|
3508
|
+
prompt?: string;
|
|
3509
|
+
max_tokens?: number;
|
|
3510
|
+
temperature?: number;
|
|
3511
|
+
ignore_eos?: boolean;
|
|
3512
|
+
top_p?: number;
|
|
3513
|
+
top_k?: number;
|
|
3514
|
+
seed?: number;
|
|
3515
|
+
repetition_penalty?: number;
|
|
3516
|
+
frequency_penalty?: number;
|
|
3517
|
+
presence_penalty?: number;
|
|
3518
|
+
raw?: boolean;
|
|
3519
|
+
messages?: RoleScopedChatInput[];
|
|
3520
|
+
};
|
|
3521
|
+
export type AiImageTextToTextOutput = {
|
|
3522
|
+
description: string;
|
|
3523
|
+
};
|
|
3524
|
+
export declare abstract class BaseAiImageTextToText {
|
|
3525
|
+
inputs: AiImageTextToTextInput;
|
|
3526
|
+
postProcessedOutputs: AiImageTextToTextOutput;
|
|
3527
|
+
}
|
|
3506
3528
|
export type AiObjectDetectionInput = {
|
|
3507
3529
|
image: number[];
|
|
3508
3530
|
};
|
|
@@ -3913,6 +3935,74 @@ export declare abstract class Base_Ai_Cf_Openai_Whisper_Large_V3_Turbo {
|
|
|
3913
3935
|
inputs: Ai_Cf_Openai_Whisper_Large_V3_Turbo_Input;
|
|
3914
3936
|
postProcessedOutputs: Ai_Cf_Openai_Whisper_Large_V3_Turbo_Output;
|
|
3915
3937
|
}
|
|
3938
|
+
export type Ai_Cf_Baai_Bge_M3_Input =
|
|
3939
|
+
| BGEM3InputQueryAndContexts
|
|
3940
|
+
| BGEM3InputEmbedding;
|
|
3941
|
+
export interface BGEM3InputQueryAndContexts {
|
|
3942
|
+
/**
|
|
3943
|
+
* A query you wish to perform against the provided contexts. If no query is provided the model with respond with embeddings for contexts
|
|
3944
|
+
*/
|
|
3945
|
+
query?: string;
|
|
3946
|
+
/**
|
|
3947
|
+
* List of provided contexts. Note that the index in this array is important, as the response will refer to it.
|
|
3948
|
+
*/
|
|
3949
|
+
contexts: {
|
|
3950
|
+
/**
|
|
3951
|
+
* One of the provided context content
|
|
3952
|
+
*/
|
|
3953
|
+
text?: string;
|
|
3954
|
+
}[];
|
|
3955
|
+
/**
|
|
3956
|
+
* When provided with too long context should the model error out or truncate the context to fit?
|
|
3957
|
+
*/
|
|
3958
|
+
truncate_inputs?: boolean;
|
|
3959
|
+
}
|
|
3960
|
+
export interface BGEM3InputEmbedding {
|
|
3961
|
+
text: string | string[];
|
|
3962
|
+
/**
|
|
3963
|
+
* When provided with too long context should the model error out or truncate the context to fit?
|
|
3964
|
+
*/
|
|
3965
|
+
truncate_inputs?: boolean;
|
|
3966
|
+
}
|
|
3967
|
+
export type Ai_Cf_Baai_Bge_M3_Output =
|
|
3968
|
+
| BGEM3OuputQuery
|
|
3969
|
+
| BGEM3OutputEmbeddingForContexts
|
|
3970
|
+
| BGEM3OuputEmbedding;
|
|
3971
|
+
export interface BGEM3OuputQuery {
|
|
3972
|
+
response?: {
|
|
3973
|
+
/**
|
|
3974
|
+
* Index of the context in the request
|
|
3975
|
+
*/
|
|
3976
|
+
id?: number;
|
|
3977
|
+
/**
|
|
3978
|
+
* Score of the context under the index.
|
|
3979
|
+
*/
|
|
3980
|
+
score?: number;
|
|
3981
|
+
}[];
|
|
3982
|
+
}
|
|
3983
|
+
export interface BGEM3OutputEmbeddingForContexts {
|
|
3984
|
+
response?: number[][];
|
|
3985
|
+
shape?: number[];
|
|
3986
|
+
/**
|
|
3987
|
+
* The pooling method used in the embedding process.
|
|
3988
|
+
*/
|
|
3989
|
+
pooling?: "mean" | "cls";
|
|
3990
|
+
}
|
|
3991
|
+
export interface BGEM3OuputEmbedding {
|
|
3992
|
+
shape?: number[];
|
|
3993
|
+
/**
|
|
3994
|
+
* Embeddings of the requested text values
|
|
3995
|
+
*/
|
|
3996
|
+
data?: number[][];
|
|
3997
|
+
/**
|
|
3998
|
+
* The pooling method used in the embedding process.
|
|
3999
|
+
*/
|
|
4000
|
+
pooling?: "mean" | "cls";
|
|
4001
|
+
}
|
|
4002
|
+
export declare abstract class Base_Ai_Cf_Baai_Bge_M3 {
|
|
4003
|
+
inputs: Ai_Cf_Baai_Bge_M3_Input;
|
|
4004
|
+
postProcessedOutputs: Ai_Cf_Baai_Bge_M3_Output;
|
|
4005
|
+
}
|
|
3916
4006
|
export interface Ai_Cf_Black_Forest_Labs_Flux_1_Schnell_Input {
|
|
3917
4007
|
/**
|
|
3918
4008
|
* A text description of the image you want to generate.
|
|
@@ -4223,6 +4313,40 @@ export declare abstract class Base_Ai_Cf_Meta_Llama_Guard_3_8B {
|
|
|
4223
4313
|
inputs: Ai_Cf_Meta_Llama_Guard_3_8B_Input;
|
|
4224
4314
|
postProcessedOutputs: Ai_Cf_Meta_Llama_Guard_3_8B_Output;
|
|
4225
4315
|
}
|
|
4316
|
+
export interface Ai_Cf_Baai_Bge_Reranker_Base_Input {
|
|
4317
|
+
/**
|
|
4318
|
+
* A query you wish to perform against the provided contexts.
|
|
4319
|
+
*/
|
|
4320
|
+
/**
|
|
4321
|
+
* Number of returned results starting with the best score.
|
|
4322
|
+
*/
|
|
4323
|
+
top_k?: number;
|
|
4324
|
+
/**
|
|
4325
|
+
* List of provided contexts. Note that the index in this array is important, as the response will refer to it.
|
|
4326
|
+
*/
|
|
4327
|
+
contexts: {
|
|
4328
|
+
/**
|
|
4329
|
+
* One of the provided context content
|
|
4330
|
+
*/
|
|
4331
|
+
text?: string;
|
|
4332
|
+
}[];
|
|
4333
|
+
}
|
|
4334
|
+
export interface Ai_Cf_Baai_Bge_Reranker_Base_Output {
|
|
4335
|
+
response?: {
|
|
4336
|
+
/**
|
|
4337
|
+
* Index of the context in the request
|
|
4338
|
+
*/
|
|
4339
|
+
id?: number;
|
|
4340
|
+
/**
|
|
4341
|
+
* Score of the context under the index.
|
|
4342
|
+
*/
|
|
4343
|
+
score?: number;
|
|
4344
|
+
}[];
|
|
4345
|
+
}
|
|
4346
|
+
export declare abstract class Base_Ai_Cf_Baai_Bge_Reranker_Base {
|
|
4347
|
+
inputs: Ai_Cf_Baai_Bge_Reranker_Base_Input;
|
|
4348
|
+
postProcessedOutputs: Ai_Cf_Baai_Bge_Reranker_Base_Output;
|
|
4349
|
+
}
|
|
4226
4350
|
export interface AiModels {
|
|
4227
4351
|
"@cf/huggingface/distilbert-sst-2-int8": BaseAiTextClassification;
|
|
4228
4352
|
"@cf/stabilityai/stable-diffusion-xl-base-1.0": BaseAiTextToImage;
|
|
@@ -4230,6 +4354,7 @@ export interface AiModels {
|
|
|
4230
4354
|
"@cf/runwayml/stable-diffusion-v1-5-img2img": BaseAiTextToImage;
|
|
4231
4355
|
"@cf/lykon/dreamshaper-8-lcm": BaseAiTextToImage;
|
|
4232
4356
|
"@cf/bytedance/stable-diffusion-xl-lightning": BaseAiTextToImage;
|
|
4357
|
+
"@cf/myshell-ai/melotts": BaseAiTextToSpeech;
|
|
4233
4358
|
"@cf/baai/bge-base-en-v1.5": BaseAiTextEmbeddings;
|
|
4234
4359
|
"@cf/baai/bge-small-en-v1.5": BaseAiTextEmbeddings;
|
|
4235
4360
|
"@cf/baai/bge-large-en-v1.5": BaseAiTextEmbeddings;
|
|
@@ -4283,9 +4408,11 @@ export interface AiModels {
|
|
|
4283
4408
|
"@cf/unum/uform-gen2-qwen-500m": Base_Ai_Cf_Unum_Uform_Gen2_Qwen_500M;
|
|
4284
4409
|
"@cf/openai/whisper-tiny-en": Base_Ai_Cf_Openai_Whisper_Tiny_En;
|
|
4285
4410
|
"@cf/openai/whisper-large-v3-turbo": Base_Ai_Cf_Openai_Whisper_Large_V3_Turbo;
|
|
4411
|
+
"@cf/baai/bge-m3": Base_Ai_Cf_Baai_Bge_M3;
|
|
4286
4412
|
"@cf/black-forest-labs/flux-1-schnell": Base_Ai_Cf_Black_Forest_Labs_Flux_1_Schnell;
|
|
4287
4413
|
"@cf/meta/llama-3.2-11b-vision-instruct": Base_Ai_Cf_Meta_Llama_3_2_11B_Vision_Instruct;
|
|
4288
4414
|
"@cf/meta/llama-guard-3-8b": Base_Ai_Cf_Meta_Llama_Guard_3_8B;
|
|
4415
|
+
"@cf/baai/bge-reranker-base": Base_Ai_Cf_Baai_Bge_Reranker_Base;
|
|
4289
4416
|
}
|
|
4290
4417
|
export type AiOptions = {
|
|
4291
4418
|
gateway?: GatewayOptions;
|
|
@@ -4345,8 +4472,8 @@ export declare abstract class Ai<
|
|
|
4345
4472
|
? Response
|
|
4346
4473
|
: AiModelList[Name]["postProcessedOutputs"]
|
|
4347
4474
|
>;
|
|
4348
|
-
|
|
4349
|
-
|
|
4475
|
+
models(params?: AiModelsSearchParams): Promise<AiModelsSearchObject[]>;
|
|
4476
|
+
toMarkdown(
|
|
4350
4477
|
files: {
|
|
4351
4478
|
name: string;
|
|
4352
4479
|
blob: Blob;
|
|
@@ -4356,7 +4483,7 @@ export declare abstract class Ai<
|
|
|
4356
4483
|
extraHeaders?: object;
|
|
4357
4484
|
},
|
|
4358
4485
|
): Promise<ConversionResponse[]>;
|
|
4359
|
-
|
|
4486
|
+
toMarkdown(
|
|
4360
4487
|
files: {
|
|
4361
4488
|
name: string;
|
|
4362
4489
|
blob: Blob;
|
package/2022-01-31/index.d.ts
CHANGED
|
@@ -3517,6 +3517,28 @@ declare abstract class BaseAiImageToText {
|
|
|
3517
3517
|
inputs: AiImageToTextInput;
|
|
3518
3518
|
postProcessedOutputs: AiImageToTextOutput;
|
|
3519
3519
|
}
|
|
3520
|
+
type AiImageTextToTextInput = {
|
|
3521
|
+
image: string;
|
|
3522
|
+
prompt?: string;
|
|
3523
|
+
max_tokens?: number;
|
|
3524
|
+
temperature?: number;
|
|
3525
|
+
ignore_eos?: boolean;
|
|
3526
|
+
top_p?: number;
|
|
3527
|
+
top_k?: number;
|
|
3528
|
+
seed?: number;
|
|
3529
|
+
repetition_penalty?: number;
|
|
3530
|
+
frequency_penalty?: number;
|
|
3531
|
+
presence_penalty?: number;
|
|
3532
|
+
raw?: boolean;
|
|
3533
|
+
messages?: RoleScopedChatInput[];
|
|
3534
|
+
};
|
|
3535
|
+
type AiImageTextToTextOutput = {
|
|
3536
|
+
description: string;
|
|
3537
|
+
};
|
|
3538
|
+
declare abstract class BaseAiImageTextToText {
|
|
3539
|
+
inputs: AiImageTextToTextInput;
|
|
3540
|
+
postProcessedOutputs: AiImageTextToTextOutput;
|
|
3541
|
+
}
|
|
3520
3542
|
type AiObjectDetectionInput = {
|
|
3521
3543
|
image: number[];
|
|
3522
3544
|
};
|
|
@@ -3927,6 +3949,72 @@ declare abstract class Base_Ai_Cf_Openai_Whisper_Large_V3_Turbo {
|
|
|
3927
3949
|
inputs: Ai_Cf_Openai_Whisper_Large_V3_Turbo_Input;
|
|
3928
3950
|
postProcessedOutputs: Ai_Cf_Openai_Whisper_Large_V3_Turbo_Output;
|
|
3929
3951
|
}
|
|
3952
|
+
type Ai_Cf_Baai_Bge_M3_Input = BGEM3InputQueryAndContexts | BGEM3InputEmbedding;
|
|
3953
|
+
interface BGEM3InputQueryAndContexts {
|
|
3954
|
+
/**
|
|
3955
|
+
* A query you wish to perform against the provided contexts. If no query is provided the model with respond with embeddings for contexts
|
|
3956
|
+
*/
|
|
3957
|
+
query?: string;
|
|
3958
|
+
/**
|
|
3959
|
+
* List of provided contexts. Note that the index in this array is important, as the response will refer to it.
|
|
3960
|
+
*/
|
|
3961
|
+
contexts: {
|
|
3962
|
+
/**
|
|
3963
|
+
* One of the provided context content
|
|
3964
|
+
*/
|
|
3965
|
+
text?: string;
|
|
3966
|
+
}[];
|
|
3967
|
+
/**
|
|
3968
|
+
* When provided with too long context should the model error out or truncate the context to fit?
|
|
3969
|
+
*/
|
|
3970
|
+
truncate_inputs?: boolean;
|
|
3971
|
+
}
|
|
3972
|
+
interface BGEM3InputEmbedding {
|
|
3973
|
+
text: string | string[];
|
|
3974
|
+
/**
|
|
3975
|
+
* When provided with too long context should the model error out or truncate the context to fit?
|
|
3976
|
+
*/
|
|
3977
|
+
truncate_inputs?: boolean;
|
|
3978
|
+
}
|
|
3979
|
+
type Ai_Cf_Baai_Bge_M3_Output =
|
|
3980
|
+
| BGEM3OuputQuery
|
|
3981
|
+
| BGEM3OutputEmbeddingForContexts
|
|
3982
|
+
| BGEM3OuputEmbedding;
|
|
3983
|
+
interface BGEM3OuputQuery {
|
|
3984
|
+
response?: {
|
|
3985
|
+
/**
|
|
3986
|
+
* Index of the context in the request
|
|
3987
|
+
*/
|
|
3988
|
+
id?: number;
|
|
3989
|
+
/**
|
|
3990
|
+
* Score of the context under the index.
|
|
3991
|
+
*/
|
|
3992
|
+
score?: number;
|
|
3993
|
+
}[];
|
|
3994
|
+
}
|
|
3995
|
+
interface BGEM3OutputEmbeddingForContexts {
|
|
3996
|
+
response?: number[][];
|
|
3997
|
+
shape?: number[];
|
|
3998
|
+
/**
|
|
3999
|
+
* The pooling method used in the embedding process.
|
|
4000
|
+
*/
|
|
4001
|
+
pooling?: "mean" | "cls";
|
|
4002
|
+
}
|
|
4003
|
+
interface BGEM3OuputEmbedding {
|
|
4004
|
+
shape?: number[];
|
|
4005
|
+
/**
|
|
4006
|
+
* Embeddings of the requested text values
|
|
4007
|
+
*/
|
|
4008
|
+
data?: number[][];
|
|
4009
|
+
/**
|
|
4010
|
+
* The pooling method used in the embedding process.
|
|
4011
|
+
*/
|
|
4012
|
+
pooling?: "mean" | "cls";
|
|
4013
|
+
}
|
|
4014
|
+
declare abstract class Base_Ai_Cf_Baai_Bge_M3 {
|
|
4015
|
+
inputs: Ai_Cf_Baai_Bge_M3_Input;
|
|
4016
|
+
postProcessedOutputs: Ai_Cf_Baai_Bge_M3_Output;
|
|
4017
|
+
}
|
|
3930
4018
|
interface Ai_Cf_Black_Forest_Labs_Flux_1_Schnell_Input {
|
|
3931
4019
|
/**
|
|
3932
4020
|
* A text description of the image you want to generate.
|
|
@@ -4237,6 +4325,40 @@ declare abstract class Base_Ai_Cf_Meta_Llama_Guard_3_8B {
|
|
|
4237
4325
|
inputs: Ai_Cf_Meta_Llama_Guard_3_8B_Input;
|
|
4238
4326
|
postProcessedOutputs: Ai_Cf_Meta_Llama_Guard_3_8B_Output;
|
|
4239
4327
|
}
|
|
4328
|
+
interface Ai_Cf_Baai_Bge_Reranker_Base_Input {
|
|
4329
|
+
/**
|
|
4330
|
+
* A query you wish to perform against the provided contexts.
|
|
4331
|
+
*/
|
|
4332
|
+
/**
|
|
4333
|
+
* Number of returned results starting with the best score.
|
|
4334
|
+
*/
|
|
4335
|
+
top_k?: number;
|
|
4336
|
+
/**
|
|
4337
|
+
* List of provided contexts. Note that the index in this array is important, as the response will refer to it.
|
|
4338
|
+
*/
|
|
4339
|
+
contexts: {
|
|
4340
|
+
/**
|
|
4341
|
+
* One of the provided context content
|
|
4342
|
+
*/
|
|
4343
|
+
text?: string;
|
|
4344
|
+
}[];
|
|
4345
|
+
}
|
|
4346
|
+
interface Ai_Cf_Baai_Bge_Reranker_Base_Output {
|
|
4347
|
+
response?: {
|
|
4348
|
+
/**
|
|
4349
|
+
* Index of the context in the request
|
|
4350
|
+
*/
|
|
4351
|
+
id?: number;
|
|
4352
|
+
/**
|
|
4353
|
+
* Score of the context under the index.
|
|
4354
|
+
*/
|
|
4355
|
+
score?: number;
|
|
4356
|
+
}[];
|
|
4357
|
+
}
|
|
4358
|
+
declare abstract class Base_Ai_Cf_Baai_Bge_Reranker_Base {
|
|
4359
|
+
inputs: Ai_Cf_Baai_Bge_Reranker_Base_Input;
|
|
4360
|
+
postProcessedOutputs: Ai_Cf_Baai_Bge_Reranker_Base_Output;
|
|
4361
|
+
}
|
|
4240
4362
|
interface AiModels {
|
|
4241
4363
|
"@cf/huggingface/distilbert-sst-2-int8": BaseAiTextClassification;
|
|
4242
4364
|
"@cf/stabilityai/stable-diffusion-xl-base-1.0": BaseAiTextToImage;
|
|
@@ -4244,6 +4366,7 @@ interface AiModels {
|
|
|
4244
4366
|
"@cf/runwayml/stable-diffusion-v1-5-img2img": BaseAiTextToImage;
|
|
4245
4367
|
"@cf/lykon/dreamshaper-8-lcm": BaseAiTextToImage;
|
|
4246
4368
|
"@cf/bytedance/stable-diffusion-xl-lightning": BaseAiTextToImage;
|
|
4369
|
+
"@cf/myshell-ai/melotts": BaseAiTextToSpeech;
|
|
4247
4370
|
"@cf/baai/bge-base-en-v1.5": BaseAiTextEmbeddings;
|
|
4248
4371
|
"@cf/baai/bge-small-en-v1.5": BaseAiTextEmbeddings;
|
|
4249
4372
|
"@cf/baai/bge-large-en-v1.5": BaseAiTextEmbeddings;
|
|
@@ -4297,9 +4420,11 @@ interface AiModels {
|
|
|
4297
4420
|
"@cf/unum/uform-gen2-qwen-500m": Base_Ai_Cf_Unum_Uform_Gen2_Qwen_500M;
|
|
4298
4421
|
"@cf/openai/whisper-tiny-en": Base_Ai_Cf_Openai_Whisper_Tiny_En;
|
|
4299
4422
|
"@cf/openai/whisper-large-v3-turbo": Base_Ai_Cf_Openai_Whisper_Large_V3_Turbo;
|
|
4423
|
+
"@cf/baai/bge-m3": Base_Ai_Cf_Baai_Bge_M3;
|
|
4300
4424
|
"@cf/black-forest-labs/flux-1-schnell": Base_Ai_Cf_Black_Forest_Labs_Flux_1_Schnell;
|
|
4301
4425
|
"@cf/meta/llama-3.2-11b-vision-instruct": Base_Ai_Cf_Meta_Llama_3_2_11B_Vision_Instruct;
|
|
4302
4426
|
"@cf/meta/llama-guard-3-8b": Base_Ai_Cf_Meta_Llama_Guard_3_8B;
|
|
4427
|
+
"@cf/baai/bge-reranker-base": Base_Ai_Cf_Baai_Bge_Reranker_Base;
|
|
4303
4428
|
}
|
|
4304
4429
|
type AiOptions = {
|
|
4305
4430
|
gateway?: GatewayOptions;
|
|
@@ -4357,8 +4482,8 @@ declare abstract class Ai<AiModelList extends AiModelListType = AiModels> {
|
|
|
4357
4482
|
? Response
|
|
4358
4483
|
: AiModelList[Name]["postProcessedOutputs"]
|
|
4359
4484
|
>;
|
|
4360
|
-
|
|
4361
|
-
|
|
4485
|
+
models(params?: AiModelsSearchParams): Promise<AiModelsSearchObject[]>;
|
|
4486
|
+
toMarkdown(
|
|
4362
4487
|
files: {
|
|
4363
4488
|
name: string;
|
|
4364
4489
|
blob: Blob;
|
|
@@ -4368,7 +4493,7 @@ declare abstract class Ai<AiModelList extends AiModelListType = AiModels> {
|
|
|
4368
4493
|
extraHeaders?: object;
|
|
4369
4494
|
},
|
|
4370
4495
|
): Promise<ConversionResponse[]>;
|
|
4371
|
-
|
|
4496
|
+
toMarkdown(
|
|
4372
4497
|
files: {
|
|
4373
4498
|
name: string;
|
|
4374
4499
|
blob: Blob;
|