@cloudflare/workers-types 4.20250402.0 → 4.20250403.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3491,6 +3491,28 @@ declare abstract class BaseAiImageToText {
3491
3491
  inputs: AiImageToTextInput;
3492
3492
  postProcessedOutputs: AiImageToTextOutput;
3493
3493
  }
3494
+ type AiImageTextToTextInput = {
3495
+ image: string;
3496
+ prompt?: string;
3497
+ max_tokens?: number;
3498
+ temperature?: number;
3499
+ ignore_eos?: boolean;
3500
+ top_p?: number;
3501
+ top_k?: number;
3502
+ seed?: number;
3503
+ repetition_penalty?: number;
3504
+ frequency_penalty?: number;
3505
+ presence_penalty?: number;
3506
+ raw?: boolean;
3507
+ messages?: RoleScopedChatInput[];
3508
+ };
3509
+ type AiImageTextToTextOutput = {
3510
+ description: string;
3511
+ };
3512
+ declare abstract class BaseAiImageTextToText {
3513
+ inputs: AiImageTextToTextInput;
3514
+ postProcessedOutputs: AiImageTextToTextOutput;
3515
+ }
3494
3516
  type AiObjectDetectionInput = {
3495
3517
  image: number[];
3496
3518
  };
@@ -3901,6 +3923,72 @@ declare abstract class Base_Ai_Cf_Openai_Whisper_Large_V3_Turbo {
3901
3923
  inputs: Ai_Cf_Openai_Whisper_Large_V3_Turbo_Input;
3902
3924
  postProcessedOutputs: Ai_Cf_Openai_Whisper_Large_V3_Turbo_Output;
3903
3925
  }
3926
+ type Ai_Cf_Baai_Bge_M3_Input = BGEM3InputQueryAndContexts | BGEM3InputEmbedding;
3927
+ interface BGEM3InputQueryAndContexts {
3928
+ /**
3929
+ * A query you wish to perform against the provided contexts. If no query is provided the model with respond with embeddings for contexts
3930
+ */
3931
+ query?: string;
3932
+ /**
3933
+ * List of provided contexts. Note that the index in this array is important, as the response will refer to it.
3934
+ */
3935
+ contexts: {
3936
+ /**
3937
+ * One of the provided context content
3938
+ */
3939
+ text?: string;
3940
+ }[];
3941
+ /**
3942
+ * When provided with too long context should the model error out or truncate the context to fit?
3943
+ */
3944
+ truncate_inputs?: boolean;
3945
+ }
3946
+ interface BGEM3InputEmbedding {
3947
+ text: string | string[];
3948
+ /**
3949
+ * When provided with too long context should the model error out or truncate the context to fit?
3950
+ */
3951
+ truncate_inputs?: boolean;
3952
+ }
3953
+ type Ai_Cf_Baai_Bge_M3_Output =
3954
+ | BGEM3OuputQuery
3955
+ | BGEM3OutputEmbeddingForContexts
3956
+ | BGEM3OuputEmbedding;
3957
+ interface BGEM3OuputQuery {
3958
+ response?: {
3959
+ /**
3960
+ * Index of the context in the request
3961
+ */
3962
+ id?: number;
3963
+ /**
3964
+ * Score of the context under the index.
3965
+ */
3966
+ score?: number;
3967
+ }[];
3968
+ }
3969
+ interface BGEM3OutputEmbeddingForContexts {
3970
+ response?: number[][];
3971
+ shape?: number[];
3972
+ /**
3973
+ * The pooling method used in the embedding process.
3974
+ */
3975
+ pooling?: "mean" | "cls";
3976
+ }
3977
+ interface BGEM3OuputEmbedding {
3978
+ shape?: number[];
3979
+ /**
3980
+ * Embeddings of the requested text values
3981
+ */
3982
+ data?: number[][];
3983
+ /**
3984
+ * The pooling method used in the embedding process.
3985
+ */
3986
+ pooling?: "mean" | "cls";
3987
+ }
3988
+ declare abstract class Base_Ai_Cf_Baai_Bge_M3 {
3989
+ inputs: Ai_Cf_Baai_Bge_M3_Input;
3990
+ postProcessedOutputs: Ai_Cf_Baai_Bge_M3_Output;
3991
+ }
3904
3992
  interface Ai_Cf_Black_Forest_Labs_Flux_1_Schnell_Input {
3905
3993
  /**
3906
3994
  * A text description of the image you want to generate.
@@ -4211,6 +4299,40 @@ declare abstract class Base_Ai_Cf_Meta_Llama_Guard_3_8B {
4211
4299
  inputs: Ai_Cf_Meta_Llama_Guard_3_8B_Input;
4212
4300
  postProcessedOutputs: Ai_Cf_Meta_Llama_Guard_3_8B_Output;
4213
4301
  }
4302
+ interface Ai_Cf_Baai_Bge_Reranker_Base_Input {
4303
+ /**
4304
+ * A query you wish to perform against the provided contexts.
4305
+ */
4306
+ /**
4307
+ * Number of returned results starting with the best score.
4308
+ */
4309
+ top_k?: number;
4310
+ /**
4311
+ * List of provided contexts. Note that the index in this array is important, as the response will refer to it.
4312
+ */
4313
+ contexts: {
4314
+ /**
4315
+ * One of the provided context content
4316
+ */
4317
+ text?: string;
4318
+ }[];
4319
+ }
4320
+ interface Ai_Cf_Baai_Bge_Reranker_Base_Output {
4321
+ response?: {
4322
+ /**
4323
+ * Index of the context in the request
4324
+ */
4325
+ id?: number;
4326
+ /**
4327
+ * Score of the context under the index.
4328
+ */
4329
+ score?: number;
4330
+ }[];
4331
+ }
4332
+ declare abstract class Base_Ai_Cf_Baai_Bge_Reranker_Base {
4333
+ inputs: Ai_Cf_Baai_Bge_Reranker_Base_Input;
4334
+ postProcessedOutputs: Ai_Cf_Baai_Bge_Reranker_Base_Output;
4335
+ }
4214
4336
  interface AiModels {
4215
4337
  "@cf/huggingface/distilbert-sst-2-int8": BaseAiTextClassification;
4216
4338
  "@cf/stabilityai/stable-diffusion-xl-base-1.0": BaseAiTextToImage;
@@ -4218,6 +4340,7 @@ interface AiModels {
4218
4340
  "@cf/runwayml/stable-diffusion-v1-5-img2img": BaseAiTextToImage;
4219
4341
  "@cf/lykon/dreamshaper-8-lcm": BaseAiTextToImage;
4220
4342
  "@cf/bytedance/stable-diffusion-xl-lightning": BaseAiTextToImage;
4343
+ "@cf/myshell-ai/melotts": BaseAiTextToSpeech;
4221
4344
  "@cf/baai/bge-base-en-v1.5": BaseAiTextEmbeddings;
4222
4345
  "@cf/baai/bge-small-en-v1.5": BaseAiTextEmbeddings;
4223
4346
  "@cf/baai/bge-large-en-v1.5": BaseAiTextEmbeddings;
@@ -4271,9 +4394,11 @@ interface AiModels {
4271
4394
  "@cf/unum/uform-gen2-qwen-500m": Base_Ai_Cf_Unum_Uform_Gen2_Qwen_500M;
4272
4395
  "@cf/openai/whisper-tiny-en": Base_Ai_Cf_Openai_Whisper_Tiny_En;
4273
4396
  "@cf/openai/whisper-large-v3-turbo": Base_Ai_Cf_Openai_Whisper_Large_V3_Turbo;
4397
+ "@cf/baai/bge-m3": Base_Ai_Cf_Baai_Bge_M3;
4274
4398
  "@cf/black-forest-labs/flux-1-schnell": Base_Ai_Cf_Black_Forest_Labs_Flux_1_Schnell;
4275
4399
  "@cf/meta/llama-3.2-11b-vision-instruct": Base_Ai_Cf_Meta_Llama_3_2_11B_Vision_Instruct;
4276
4400
  "@cf/meta/llama-guard-3-8b": Base_Ai_Cf_Meta_Llama_Guard_3_8B;
4401
+ "@cf/baai/bge-reranker-base": Base_Ai_Cf_Baai_Bge_Reranker_Base;
4277
4402
  }
4278
4403
  type AiOptions = {
4279
4404
  gateway?: GatewayOptions;
@@ -4331,8 +4456,8 @@ declare abstract class Ai<AiModelList extends AiModelListType = AiModels> {
4331
4456
  ? Response
4332
4457
  : AiModelList[Name]["postProcessedOutputs"]
4333
4458
  >;
4334
- public models(params?: AiModelsSearchParams): Promise<AiModelsSearchObject[]>;
4335
- public toMarkdown(
4459
+ models(params?: AiModelsSearchParams): Promise<AiModelsSearchObject[]>;
4460
+ toMarkdown(
4336
4461
  files: {
4337
4462
  name: string;
4338
4463
  blob: Blob;
@@ -4342,7 +4467,7 @@ declare abstract class Ai<AiModelList extends AiModelListType = AiModels> {
4342
4467
  extraHeaders?: object;
4343
4468
  },
4344
4469
  ): Promise<ConversionResponse[]>;
4345
- public toMarkdown(
4470
+ toMarkdown(
4346
4471
  files: {
4347
4472
  name: string;
4348
4473
  blob: Blob;
@@ -3503,6 +3503,28 @@ export declare abstract class BaseAiImageToText {
3503
3503
  inputs: AiImageToTextInput;
3504
3504
  postProcessedOutputs: AiImageToTextOutput;
3505
3505
  }
3506
+ export type AiImageTextToTextInput = {
3507
+ image: string;
3508
+ prompt?: string;
3509
+ max_tokens?: number;
3510
+ temperature?: number;
3511
+ ignore_eos?: boolean;
3512
+ top_p?: number;
3513
+ top_k?: number;
3514
+ seed?: number;
3515
+ repetition_penalty?: number;
3516
+ frequency_penalty?: number;
3517
+ presence_penalty?: number;
3518
+ raw?: boolean;
3519
+ messages?: RoleScopedChatInput[];
3520
+ };
3521
+ export type AiImageTextToTextOutput = {
3522
+ description: string;
3523
+ };
3524
+ export declare abstract class BaseAiImageTextToText {
3525
+ inputs: AiImageTextToTextInput;
3526
+ postProcessedOutputs: AiImageTextToTextOutput;
3527
+ }
3506
3528
  export type AiObjectDetectionInput = {
3507
3529
  image: number[];
3508
3530
  };
@@ -3913,6 +3935,74 @@ export declare abstract class Base_Ai_Cf_Openai_Whisper_Large_V3_Turbo {
3913
3935
  inputs: Ai_Cf_Openai_Whisper_Large_V3_Turbo_Input;
3914
3936
  postProcessedOutputs: Ai_Cf_Openai_Whisper_Large_V3_Turbo_Output;
3915
3937
  }
3938
+ export type Ai_Cf_Baai_Bge_M3_Input =
3939
+ | BGEM3InputQueryAndContexts
3940
+ | BGEM3InputEmbedding;
3941
+ export interface BGEM3InputQueryAndContexts {
3942
+ /**
3943
+ * A query you wish to perform against the provided contexts. If no query is provided the model with respond with embeddings for contexts
3944
+ */
3945
+ query?: string;
3946
+ /**
3947
+ * List of provided contexts. Note that the index in this array is important, as the response will refer to it.
3948
+ */
3949
+ contexts: {
3950
+ /**
3951
+ * One of the provided context content
3952
+ */
3953
+ text?: string;
3954
+ }[];
3955
+ /**
3956
+ * When provided with too long context should the model error out or truncate the context to fit?
3957
+ */
3958
+ truncate_inputs?: boolean;
3959
+ }
3960
+ export interface BGEM3InputEmbedding {
3961
+ text: string | string[];
3962
+ /**
3963
+ * When provided with too long context should the model error out or truncate the context to fit?
3964
+ */
3965
+ truncate_inputs?: boolean;
3966
+ }
3967
+ export type Ai_Cf_Baai_Bge_M3_Output =
3968
+ | BGEM3OuputQuery
3969
+ | BGEM3OutputEmbeddingForContexts
3970
+ | BGEM3OuputEmbedding;
3971
+ export interface BGEM3OuputQuery {
3972
+ response?: {
3973
+ /**
3974
+ * Index of the context in the request
3975
+ */
3976
+ id?: number;
3977
+ /**
3978
+ * Score of the context under the index.
3979
+ */
3980
+ score?: number;
3981
+ }[];
3982
+ }
3983
+ export interface BGEM3OutputEmbeddingForContexts {
3984
+ response?: number[][];
3985
+ shape?: number[];
3986
+ /**
3987
+ * The pooling method used in the embedding process.
3988
+ */
3989
+ pooling?: "mean" | "cls";
3990
+ }
3991
+ export interface BGEM3OuputEmbedding {
3992
+ shape?: number[];
3993
+ /**
3994
+ * Embeddings of the requested text values
3995
+ */
3996
+ data?: number[][];
3997
+ /**
3998
+ * The pooling method used in the embedding process.
3999
+ */
4000
+ pooling?: "mean" | "cls";
4001
+ }
4002
+ export declare abstract class Base_Ai_Cf_Baai_Bge_M3 {
4003
+ inputs: Ai_Cf_Baai_Bge_M3_Input;
4004
+ postProcessedOutputs: Ai_Cf_Baai_Bge_M3_Output;
4005
+ }
3916
4006
  export interface Ai_Cf_Black_Forest_Labs_Flux_1_Schnell_Input {
3917
4007
  /**
3918
4008
  * A text description of the image you want to generate.
@@ -4223,6 +4313,40 @@ export declare abstract class Base_Ai_Cf_Meta_Llama_Guard_3_8B {
4223
4313
  inputs: Ai_Cf_Meta_Llama_Guard_3_8B_Input;
4224
4314
  postProcessedOutputs: Ai_Cf_Meta_Llama_Guard_3_8B_Output;
4225
4315
  }
4316
+ export interface Ai_Cf_Baai_Bge_Reranker_Base_Input {
4317
+ /**
4318
+ * A query you wish to perform against the provided contexts.
4319
+ */
4320
+ /**
4321
+ * Number of returned results starting with the best score.
4322
+ */
4323
+ top_k?: number;
4324
+ /**
4325
+ * List of provided contexts. Note that the index in this array is important, as the response will refer to it.
4326
+ */
4327
+ contexts: {
4328
+ /**
4329
+ * One of the provided context content
4330
+ */
4331
+ text?: string;
4332
+ }[];
4333
+ }
4334
+ export interface Ai_Cf_Baai_Bge_Reranker_Base_Output {
4335
+ response?: {
4336
+ /**
4337
+ * Index of the context in the request
4338
+ */
4339
+ id?: number;
4340
+ /**
4341
+ * Score of the context under the index.
4342
+ */
4343
+ score?: number;
4344
+ }[];
4345
+ }
4346
+ export declare abstract class Base_Ai_Cf_Baai_Bge_Reranker_Base {
4347
+ inputs: Ai_Cf_Baai_Bge_Reranker_Base_Input;
4348
+ postProcessedOutputs: Ai_Cf_Baai_Bge_Reranker_Base_Output;
4349
+ }
4226
4350
  export interface AiModels {
4227
4351
  "@cf/huggingface/distilbert-sst-2-int8": BaseAiTextClassification;
4228
4352
  "@cf/stabilityai/stable-diffusion-xl-base-1.0": BaseAiTextToImage;
@@ -4230,6 +4354,7 @@ export interface AiModels {
4230
4354
  "@cf/runwayml/stable-diffusion-v1-5-img2img": BaseAiTextToImage;
4231
4355
  "@cf/lykon/dreamshaper-8-lcm": BaseAiTextToImage;
4232
4356
  "@cf/bytedance/stable-diffusion-xl-lightning": BaseAiTextToImage;
4357
+ "@cf/myshell-ai/melotts": BaseAiTextToSpeech;
4233
4358
  "@cf/baai/bge-base-en-v1.5": BaseAiTextEmbeddings;
4234
4359
  "@cf/baai/bge-small-en-v1.5": BaseAiTextEmbeddings;
4235
4360
  "@cf/baai/bge-large-en-v1.5": BaseAiTextEmbeddings;
@@ -4283,9 +4408,11 @@ export interface AiModels {
4283
4408
  "@cf/unum/uform-gen2-qwen-500m": Base_Ai_Cf_Unum_Uform_Gen2_Qwen_500M;
4284
4409
  "@cf/openai/whisper-tiny-en": Base_Ai_Cf_Openai_Whisper_Tiny_En;
4285
4410
  "@cf/openai/whisper-large-v3-turbo": Base_Ai_Cf_Openai_Whisper_Large_V3_Turbo;
4411
+ "@cf/baai/bge-m3": Base_Ai_Cf_Baai_Bge_M3;
4286
4412
  "@cf/black-forest-labs/flux-1-schnell": Base_Ai_Cf_Black_Forest_Labs_Flux_1_Schnell;
4287
4413
  "@cf/meta/llama-3.2-11b-vision-instruct": Base_Ai_Cf_Meta_Llama_3_2_11B_Vision_Instruct;
4288
4414
  "@cf/meta/llama-guard-3-8b": Base_Ai_Cf_Meta_Llama_Guard_3_8B;
4415
+ "@cf/baai/bge-reranker-base": Base_Ai_Cf_Baai_Bge_Reranker_Base;
4289
4416
  }
4290
4417
  export type AiOptions = {
4291
4418
  gateway?: GatewayOptions;
@@ -4345,8 +4472,8 @@ export declare abstract class Ai<
4345
4472
  ? Response
4346
4473
  : AiModelList[Name]["postProcessedOutputs"]
4347
4474
  >;
4348
- public models(params?: AiModelsSearchParams): Promise<AiModelsSearchObject[]>;
4349
- public toMarkdown(
4475
+ models(params?: AiModelsSearchParams): Promise<AiModelsSearchObject[]>;
4476
+ toMarkdown(
4350
4477
  files: {
4351
4478
  name: string;
4352
4479
  blob: Blob;
@@ -4356,7 +4483,7 @@ export declare abstract class Ai<
4356
4483
  extraHeaders?: object;
4357
4484
  },
4358
4485
  ): Promise<ConversionResponse[]>;
4359
- public toMarkdown(
4486
+ toMarkdown(
4360
4487
  files: {
4361
4488
  name: string;
4362
4489
  blob: Blob;
@@ -3517,6 +3517,28 @@ declare abstract class BaseAiImageToText {
3517
3517
  inputs: AiImageToTextInput;
3518
3518
  postProcessedOutputs: AiImageToTextOutput;
3519
3519
  }
3520
+ type AiImageTextToTextInput = {
3521
+ image: string;
3522
+ prompt?: string;
3523
+ max_tokens?: number;
3524
+ temperature?: number;
3525
+ ignore_eos?: boolean;
3526
+ top_p?: number;
3527
+ top_k?: number;
3528
+ seed?: number;
3529
+ repetition_penalty?: number;
3530
+ frequency_penalty?: number;
3531
+ presence_penalty?: number;
3532
+ raw?: boolean;
3533
+ messages?: RoleScopedChatInput[];
3534
+ };
3535
+ type AiImageTextToTextOutput = {
3536
+ description: string;
3537
+ };
3538
+ declare abstract class BaseAiImageTextToText {
3539
+ inputs: AiImageTextToTextInput;
3540
+ postProcessedOutputs: AiImageTextToTextOutput;
3541
+ }
3520
3542
  type AiObjectDetectionInput = {
3521
3543
  image: number[];
3522
3544
  };
@@ -3927,6 +3949,72 @@ declare abstract class Base_Ai_Cf_Openai_Whisper_Large_V3_Turbo {
3927
3949
  inputs: Ai_Cf_Openai_Whisper_Large_V3_Turbo_Input;
3928
3950
  postProcessedOutputs: Ai_Cf_Openai_Whisper_Large_V3_Turbo_Output;
3929
3951
  }
3952
+ type Ai_Cf_Baai_Bge_M3_Input = BGEM3InputQueryAndContexts | BGEM3InputEmbedding;
3953
+ interface BGEM3InputQueryAndContexts {
3954
+ /**
3955
+ * A query you wish to perform against the provided contexts. If no query is provided the model with respond with embeddings for contexts
3956
+ */
3957
+ query?: string;
3958
+ /**
3959
+ * List of provided contexts. Note that the index in this array is important, as the response will refer to it.
3960
+ */
3961
+ contexts: {
3962
+ /**
3963
+ * One of the provided context content
3964
+ */
3965
+ text?: string;
3966
+ }[];
3967
+ /**
3968
+ * When provided with too long context should the model error out or truncate the context to fit?
3969
+ */
3970
+ truncate_inputs?: boolean;
3971
+ }
3972
+ interface BGEM3InputEmbedding {
3973
+ text: string | string[];
3974
+ /**
3975
+ * When provided with too long context should the model error out or truncate the context to fit?
3976
+ */
3977
+ truncate_inputs?: boolean;
3978
+ }
3979
+ type Ai_Cf_Baai_Bge_M3_Output =
3980
+ | BGEM3OuputQuery
3981
+ | BGEM3OutputEmbeddingForContexts
3982
+ | BGEM3OuputEmbedding;
3983
+ interface BGEM3OuputQuery {
3984
+ response?: {
3985
+ /**
3986
+ * Index of the context in the request
3987
+ */
3988
+ id?: number;
3989
+ /**
3990
+ * Score of the context under the index.
3991
+ */
3992
+ score?: number;
3993
+ }[];
3994
+ }
3995
+ interface BGEM3OutputEmbeddingForContexts {
3996
+ response?: number[][];
3997
+ shape?: number[];
3998
+ /**
3999
+ * The pooling method used in the embedding process.
4000
+ */
4001
+ pooling?: "mean" | "cls";
4002
+ }
4003
+ interface BGEM3OuputEmbedding {
4004
+ shape?: number[];
4005
+ /**
4006
+ * Embeddings of the requested text values
4007
+ */
4008
+ data?: number[][];
4009
+ /**
4010
+ * The pooling method used in the embedding process.
4011
+ */
4012
+ pooling?: "mean" | "cls";
4013
+ }
4014
+ declare abstract class Base_Ai_Cf_Baai_Bge_M3 {
4015
+ inputs: Ai_Cf_Baai_Bge_M3_Input;
4016
+ postProcessedOutputs: Ai_Cf_Baai_Bge_M3_Output;
4017
+ }
3930
4018
  interface Ai_Cf_Black_Forest_Labs_Flux_1_Schnell_Input {
3931
4019
  /**
3932
4020
  * A text description of the image you want to generate.
@@ -4237,6 +4325,40 @@ declare abstract class Base_Ai_Cf_Meta_Llama_Guard_3_8B {
4237
4325
  inputs: Ai_Cf_Meta_Llama_Guard_3_8B_Input;
4238
4326
  postProcessedOutputs: Ai_Cf_Meta_Llama_Guard_3_8B_Output;
4239
4327
  }
4328
+ interface Ai_Cf_Baai_Bge_Reranker_Base_Input {
4329
+ /**
4330
+ * A query you wish to perform against the provided contexts.
4331
+ */
4332
+ /**
4333
+ * Number of returned results starting with the best score.
4334
+ */
4335
+ top_k?: number;
4336
+ /**
4337
+ * List of provided contexts. Note that the index in this array is important, as the response will refer to it.
4338
+ */
4339
+ contexts: {
4340
+ /**
4341
+ * One of the provided context content
4342
+ */
4343
+ text?: string;
4344
+ }[];
4345
+ }
4346
+ interface Ai_Cf_Baai_Bge_Reranker_Base_Output {
4347
+ response?: {
4348
+ /**
4349
+ * Index of the context in the request
4350
+ */
4351
+ id?: number;
4352
+ /**
4353
+ * Score of the context under the index.
4354
+ */
4355
+ score?: number;
4356
+ }[];
4357
+ }
4358
+ declare abstract class Base_Ai_Cf_Baai_Bge_Reranker_Base {
4359
+ inputs: Ai_Cf_Baai_Bge_Reranker_Base_Input;
4360
+ postProcessedOutputs: Ai_Cf_Baai_Bge_Reranker_Base_Output;
4361
+ }
4240
4362
  interface AiModels {
4241
4363
  "@cf/huggingface/distilbert-sst-2-int8": BaseAiTextClassification;
4242
4364
  "@cf/stabilityai/stable-diffusion-xl-base-1.0": BaseAiTextToImage;
@@ -4244,6 +4366,7 @@ interface AiModels {
4244
4366
  "@cf/runwayml/stable-diffusion-v1-5-img2img": BaseAiTextToImage;
4245
4367
  "@cf/lykon/dreamshaper-8-lcm": BaseAiTextToImage;
4246
4368
  "@cf/bytedance/stable-diffusion-xl-lightning": BaseAiTextToImage;
4369
+ "@cf/myshell-ai/melotts": BaseAiTextToSpeech;
4247
4370
  "@cf/baai/bge-base-en-v1.5": BaseAiTextEmbeddings;
4248
4371
  "@cf/baai/bge-small-en-v1.5": BaseAiTextEmbeddings;
4249
4372
  "@cf/baai/bge-large-en-v1.5": BaseAiTextEmbeddings;
@@ -4297,9 +4420,11 @@ interface AiModels {
4297
4420
  "@cf/unum/uform-gen2-qwen-500m": Base_Ai_Cf_Unum_Uform_Gen2_Qwen_500M;
4298
4421
  "@cf/openai/whisper-tiny-en": Base_Ai_Cf_Openai_Whisper_Tiny_En;
4299
4422
  "@cf/openai/whisper-large-v3-turbo": Base_Ai_Cf_Openai_Whisper_Large_V3_Turbo;
4423
+ "@cf/baai/bge-m3": Base_Ai_Cf_Baai_Bge_M3;
4300
4424
  "@cf/black-forest-labs/flux-1-schnell": Base_Ai_Cf_Black_Forest_Labs_Flux_1_Schnell;
4301
4425
  "@cf/meta/llama-3.2-11b-vision-instruct": Base_Ai_Cf_Meta_Llama_3_2_11B_Vision_Instruct;
4302
4426
  "@cf/meta/llama-guard-3-8b": Base_Ai_Cf_Meta_Llama_Guard_3_8B;
4427
+ "@cf/baai/bge-reranker-base": Base_Ai_Cf_Baai_Bge_Reranker_Base;
4303
4428
  }
4304
4429
  type AiOptions = {
4305
4430
  gateway?: GatewayOptions;
@@ -4357,8 +4482,8 @@ declare abstract class Ai<AiModelList extends AiModelListType = AiModels> {
4357
4482
  ? Response
4358
4483
  : AiModelList[Name]["postProcessedOutputs"]
4359
4484
  >;
4360
- public models(params?: AiModelsSearchParams): Promise<AiModelsSearchObject[]>;
4361
- public toMarkdown(
4485
+ models(params?: AiModelsSearchParams): Promise<AiModelsSearchObject[]>;
4486
+ toMarkdown(
4362
4487
  files: {
4363
4488
  name: string;
4364
4489
  blob: Blob;
@@ -4368,7 +4493,7 @@ declare abstract class Ai<AiModelList extends AiModelListType = AiModels> {
4368
4493
  extraHeaders?: object;
4369
4494
  },
4370
4495
  ): Promise<ConversionResponse[]>;
4371
- public toMarkdown(
4496
+ toMarkdown(
4372
4497
  files: {
4373
4498
  name: string;
4374
4499
  blob: Blob;