modelfusion 0.135.1 → 0.137.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +2359 -0
- package/README.md +697 -0
- package/index.cjs +103 -305
- package/index.cjs.map +1 -1
- package/index.d.cts +112 -197
- package/index.d.ts +112 -197
- package/index.js +101 -292
- package/index.js.map +1 -1
- package/node/index.cjs +184 -0
- package/node/index.cjs.map +1 -0
- package/node/index.d.cts +92 -0
- package/node/index.d.ts +92 -0
- package/node/index.js +147 -0
- package/node/index.js.map +1 -0
- package/package.json +11 -3
    
        package/index.d.ts
    CHANGED
    
    | @@ -1,5 +1,5 @@ | |
| 1 1 | 
             
            import { PartialDeep } from 'type-fest';
         | 
| 2 | 
            -
            import { z } from 'zod';
         | 
| 2 | 
            +
            import z$1, { z } from 'zod';
         | 
| 3 3 |  | 
| 4 4 | 
             
            type ErrorHandler = (error: unknown) => void;
         | 
| 5 5 |  | 
| @@ -463,6 +463,11 @@ declare class DefaultRun implements Run { | |
| 463 463 | 
             
                readonly functionObserver: {
         | 
| 464 464 | 
             
                    onFunctionEvent: (event: FunctionEvent) => void;
         | 
| 465 465 | 
             
                };
         | 
| 466 | 
            +
                getSuccessfulModelCalls(): (ModelCallFinishedEvent & {
         | 
| 467 | 
            +
                    result: {
         | 
| 468 | 
            +
                        status: "success";
         | 
| 469 | 
            +
                    };
         | 
| 470 | 
            +
                })[];
         | 
| 466 471 | 
             
            }
         | 
| 467 472 |  | 
| 468 473 | 
             
            declare class FunctionEventSource {
         | 
| @@ -3989,8 +3994,8 @@ declare class LlamaCppCompletionModel<CONTEXT_WINDOW_SIZE extends number | undef | |
| 3989 3994 | 
             
                countPromptTokens(prompt: LlamaCppCompletionPrompt): Promise<number>;
         | 
| 3990 3995 | 
             
                doGenerateTexts(prompt: LlamaCppCompletionPrompt, options: FunctionCallOptions): Promise<{
         | 
| 3991 3996 | 
             
                    rawResponse: {
         | 
| 3992 | 
            -
                        prompt: string;
         | 
| 3993 3997 | 
             
                        model: string;
         | 
| 3998 | 
            +
                        prompt: string;
         | 
| 3994 3999 | 
             
                        stop: true;
         | 
| 3995 4000 | 
             
                        content: string;
         | 
| 3996 4001 | 
             
                        generation_settings: {
         | 
| @@ -4049,8 +4054,8 @@ declare class LlamaCppCompletionModel<CONTEXT_WINDOW_SIZE extends number | undef | |
| 4049 4054 | 
             
                }>;
         | 
| 4050 4055 | 
             
                restoreGeneratedTexts(rawResponse: unknown): {
         | 
| 4051 4056 | 
             
                    rawResponse: {
         | 
| 4052 | 
            -
                        prompt: string;
         | 
| 4053 4057 | 
             
                        model: string;
         | 
| 4058 | 
            +
                        prompt: string;
         | 
| 4054 4059 | 
             
                        stop: true;
         | 
| 4055 4060 | 
             
                        content: string;
         | 
| 4056 4061 | 
             
                        generation_settings: {
         | 
| @@ -4109,8 +4114,8 @@ declare class LlamaCppCompletionModel<CONTEXT_WINDOW_SIZE extends number | undef | |
| 4109 4114 | 
             
                };
         | 
| 4110 4115 | 
             
                processTextGenerationResponse(rawResponse: LlamaCppTextGenerationResponse): {
         | 
| 4111 4116 | 
             
                    rawResponse: {
         | 
| 4112 | 
            -
                        prompt: string;
         | 
| 4113 4117 | 
             
                        model: string;
         | 
| 4118 | 
            +
                        prompt: string;
         | 
| 4114 4119 | 
             
                        stop: true;
         | 
| 4115 4120 | 
             
                        content: string;
         | 
| 4116 4121 | 
             
                        generation_settings: {
         | 
| @@ -4168,8 +4173,8 @@ declare class LlamaCppCompletionModel<CONTEXT_WINDOW_SIZE extends number | undef | |
| 4168 4173 | 
             
                    };
         | 
| 4169 4174 | 
             
                };
         | 
| 4170 4175 | 
             
                doStreamText(prompt: LlamaCppCompletionPrompt, options: FunctionCallOptions): Promise<AsyncIterable<Delta<{
         | 
| 4171 | 
            -
                    prompt: string;
         | 
| 4172 4176 | 
             
                    model: string;
         | 
| 4177 | 
            +
                    prompt: string;
         | 
| 4173 4178 | 
             
                    stop: true;
         | 
| 4174 4179 | 
             
                    content: string;
         | 
| 4175 4180 | 
             
                    generation_settings: {
         | 
| @@ -4347,8 +4352,8 @@ declare const llamaCppTextGenerationResponseSchema: z.ZodObject<{ | |
| 4347 4352 | 
             
                tokens_predicted: z.ZodNumber;
         | 
| 4348 4353 | 
             
                truncated: z.ZodBoolean;
         | 
| 4349 4354 | 
             
            }, "strip", z.ZodTypeAny, {
         | 
| 4350 | 
            -
                prompt: string;
         | 
| 4351 4355 | 
             
                model: string;
         | 
| 4356 | 
            +
                prompt: string;
         | 
| 4352 4357 | 
             
                stop: true;
         | 
| 4353 4358 | 
             
                content: string;
         | 
| 4354 4359 | 
             
                generation_settings: {
         | 
| @@ -4395,8 +4400,8 @@ declare const llamaCppTextGenerationResponseSchema: z.ZodObject<{ | |
| 4395 4400 | 
             
                tokens_predicted: number;
         | 
| 4396 4401 | 
             
                truncated: boolean;
         | 
| 4397 4402 | 
             
            }, {
         | 
| 4398 | 
            -
                prompt: string;
         | 
| 4399 4403 | 
             
                model: string;
         | 
| 4404 | 
            +
                prompt: string;
         | 
| 4400 4405 | 
             
                stop: true;
         | 
| 4401 4406 | 
             
                content: string;
         | 
| 4402 4407 | 
             
                generation_settings: {
         | 
| @@ -4568,8 +4573,8 @@ declare const llamaCppTextStreamChunkSchema: z.ZodDiscriminatedUnion<"stop", [z. | |
| 4568 4573 | 
             
                tokens_predicted: z.ZodNumber;
         | 
| 4569 4574 | 
             
                truncated: z.ZodBoolean;
         | 
| 4570 4575 | 
             
            }, "strip", z.ZodTypeAny, {
         | 
| 4571 | 
            -
                prompt: string;
         | 
| 4572 4576 | 
             
                model: string;
         | 
| 4577 | 
            +
                prompt: string;
         | 
| 4573 4578 | 
             
                stop: true;
         | 
| 4574 4579 | 
             
                content: string;
         | 
| 4575 4580 | 
             
                generation_settings: {
         | 
| @@ -4616,8 +4621,8 @@ declare const llamaCppTextStreamChunkSchema: z.ZodDiscriminatedUnion<"stop", [z. | |
| 4616 4621 | 
             
                tokens_predicted: number;
         | 
| 4617 4622 | 
             
                truncated: boolean;
         | 
| 4618 4623 | 
             
            }, {
         | 
| 4619 | 
            -
                prompt: string;
         | 
| 4620 4624 | 
             
                model: string;
         | 
| 4625 | 
            +
                prompt: string;
         | 
| 4621 4626 | 
             
                stop: true;
         | 
| 4622 4627 | 
             
                content: string;
         | 
| 4623 4628 | 
             
                generation_settings: {
         | 
| @@ -4676,8 +4681,8 @@ declare const LlamaCppCompletionResponseFormat: { | |
| 4676 4681 | 
             
                json: {
         | 
| 4677 4682 | 
             
                    stream: false;
         | 
| 4678 4683 | 
             
                    handler: ResponseHandler<{
         | 
| 4679 | 
            -
                        prompt: string;
         | 
| 4680 4684 | 
             
                        model: string;
         | 
| 4685 | 
            +
                        prompt: string;
         | 
| 4681 4686 | 
             
                        stop: true;
         | 
| 4682 4687 | 
             
                        content: string;
         | 
| 4683 4688 | 
             
                        generation_settings: {
         | 
| @@ -4734,8 +4739,8 @@ declare const LlamaCppCompletionResponseFormat: { | |
| 4734 4739 | 
             
                    handler: ({ response }: {
         | 
| 4735 4740 | 
             
                        response: Response;
         | 
| 4736 4741 | 
             
                    }) => Promise<AsyncIterable<Delta<{
         | 
| 4737 | 
            -
                        prompt: string;
         | 
| 4738 4742 | 
             
                        model: string;
         | 
| 4743 | 
            +
                        prompt: string;
         | 
| 4739 4744 | 
             
                        stop: true;
         | 
| 4740 4745 | 
             
                        content: string;
         | 
| 4741 4746 | 
             
                        generation_settings: {
         | 
| @@ -5454,12 +5459,12 @@ declare class MistralTextEmbeddingModel extends AbstractModel<MistralTextEmbeddi | |
| 5454 5459 | 
             
                doEmbedValues(texts: string[], options: FunctionCallOptions): Promise<{
         | 
| 5455 5460 | 
             
                    rawResponse: {
         | 
| 5456 5461 | 
             
                        object: string;
         | 
| 5462 | 
            +
                        model: string;
         | 
| 5457 5463 | 
             
                        data: {
         | 
| 5458 5464 | 
             
                            object: string;
         | 
| 5459 5465 | 
             
                            embedding: number[];
         | 
| 5460 5466 | 
             
                            index: number;
         | 
| 5461 5467 | 
             
                        }[];
         | 
| 5462 | 
            -
                        model: string;
         | 
| 5463 5468 | 
             
                        usage: {
         | 
| 5464 5469 | 
             
                            prompt_tokens: number;
         | 
| 5465 5470 | 
             
                            total_tokens: number;
         | 
| @@ -5499,12 +5504,12 @@ declare const MistralTextEmbeddingResponseSchema: z.ZodObject<{ | |
| 5499 5504 | 
             
                }>;
         | 
| 5500 5505 | 
             
            }, "strip", z.ZodTypeAny, {
         | 
| 5501 5506 | 
             
                object: string;
         | 
| 5507 | 
            +
                model: string;
         | 
| 5502 5508 | 
             
                data: {
         | 
| 5503 5509 | 
             
                    object: string;
         | 
| 5504 5510 | 
             
                    embedding: number[];
         | 
| 5505 5511 | 
             
                    index: number;
         | 
| 5506 5512 | 
             
                }[];
         | 
| 5507 | 
            -
                model: string;
         | 
| 5508 5513 | 
             
                usage: {
         | 
| 5509 5514 | 
             
                    prompt_tokens: number;
         | 
| 5510 5515 | 
             
                    total_tokens: number;
         | 
| @@ -5512,12 +5517,12 @@ declare const MistralTextEmbeddingResponseSchema: z.ZodObject<{ | |
| 5512 5517 | 
             
                id: string;
         | 
| 5513 5518 | 
             
            }, {
         | 
| 5514 5519 | 
             
                object: string;
         | 
| 5520 | 
            +
                model: string;
         | 
| 5515 5521 | 
             
                data: {
         | 
| 5516 5522 | 
             
                    object: string;
         | 
| 5517 5523 | 
             
                    embedding: number[];
         | 
| 5518 5524 | 
             
                    index: number;
         | 
| 5519 5525 | 
             
                }[];
         | 
| 5520 | 
            -
                model: string;
         | 
| 5521 5526 | 
             
                usage: {
         | 
| 5522 5527 | 
             
                    prompt_tokens: number;
         | 
| 5523 5528 | 
             
                    total_tokens: number;
         | 
| @@ -5663,11 +5668,11 @@ declare class OllamaChatModel extends AbstractModel<OllamaChatModelSettings> imp | |
| 5663 5668 | 
             
                get settingsForEvent(): Partial<OllamaChatModelSettings>;
         | 
| 5664 5669 | 
             
                doGenerateTexts(prompt: OllamaChatPrompt, options: FunctionCallOptions): Promise<{
         | 
| 5665 5670 | 
             
                    rawResponse: {
         | 
| 5671 | 
            +
                        model: string;
         | 
| 5666 5672 | 
             
                        message: {
         | 
| 5667 5673 | 
             
                            role: string;
         | 
| 5668 5674 | 
             
                            content: string;
         | 
| 5669 5675 | 
             
                        };
         | 
| 5670 | 
            -
                        model: string;
         | 
| 5671 5676 | 
             
                        done: true;
         | 
| 5672 5677 | 
             
                        created_at: string;
         | 
| 5673 5678 | 
             
                        total_duration: number;
         | 
| @@ -5684,11 +5689,11 @@ declare class OllamaChatModel extends AbstractModel<OllamaChatModelSettings> imp | |
| 5684 5689 | 
             
                }>;
         | 
| 5685 5690 | 
             
                restoreGeneratedTexts(rawResponse: unknown): {
         | 
| 5686 5691 | 
             
                    rawResponse: {
         | 
| 5692 | 
            +
                        model: string;
         | 
| 5687 5693 | 
             
                        message: {
         | 
| 5688 5694 | 
             
                            role: string;
         | 
| 5689 5695 | 
             
                            content: string;
         | 
| 5690 5696 | 
             
                        };
         | 
| 5691 | 
            -
                        model: string;
         | 
| 5692 5697 | 
             
                        done: true;
         | 
| 5693 5698 | 
             
                        created_at: string;
         | 
| 5694 5699 | 
             
                        total_duration: number;
         | 
| @@ -5705,11 +5710,11 @@ declare class OllamaChatModel extends AbstractModel<OllamaChatModelSettings> imp | |
| 5705 5710 | 
             
                };
         | 
| 5706 5711 | 
             
                private processTextGenerationResponse;
         | 
| 5707 5712 | 
             
                doStreamText(prompt: OllamaChatPrompt, options: FunctionCallOptions): Promise<AsyncIterable<Delta<{
         | 
| 5713 | 
            +
                    model: string;
         | 
| 5708 5714 | 
             
                    message: {
         | 
| 5709 5715 | 
             
                        role: string;
         | 
| 5710 5716 | 
             
                        content: string;
         | 
| 5711 5717 | 
             
                    };
         | 
| 5712 | 
            -
                    model: string;
         | 
| 5713 5718 | 
             
                    done: false;
         | 
| 5714 5719 | 
             
                    created_at: string;
         | 
| 5715 5720 | 
             
                } | {
         | 
| @@ -5755,11 +5760,11 @@ declare const ollamaChatResponseSchema: z.ZodObject<{ | |
| 5755 5760 | 
             
                eval_count: z.ZodNumber;
         | 
| 5756 5761 | 
             
                eval_duration: z.ZodNumber;
         | 
| 5757 5762 | 
             
            }, "strip", z.ZodTypeAny, {
         | 
| 5763 | 
            +
                model: string;
         | 
| 5758 5764 | 
             
                message: {
         | 
| 5759 5765 | 
             
                    role: string;
         | 
| 5760 5766 | 
             
                    content: string;
         | 
| 5761 5767 | 
             
                };
         | 
| 5762 | 
            -
                model: string;
         | 
| 5763 5768 | 
             
                done: true;
         | 
| 5764 5769 | 
             
                created_at: string;
         | 
| 5765 5770 | 
             
                total_duration: number;
         | 
| @@ -5769,11 +5774,11 @@ declare const ollamaChatResponseSchema: z.ZodObject<{ | |
| 5769 5774 | 
             
                prompt_eval_count?: number | undefined;
         | 
| 5770 5775 | 
             
                prompt_eval_duration?: number | undefined;
         | 
| 5771 5776 | 
             
            }, {
         | 
| 5777 | 
            +
                model: string;
         | 
| 5772 5778 | 
             
                message: {
         | 
| 5773 5779 | 
             
                    role: string;
         | 
| 5774 5780 | 
             
                    content: string;
         | 
| 5775 5781 | 
             
                };
         | 
| 5776 | 
            -
                model: string;
         | 
| 5777 5782 | 
             
                done: true;
         | 
| 5778 5783 | 
             
                created_at: string;
         | 
| 5779 5784 | 
             
                total_duration: number;
         | 
| @@ -5799,19 +5804,19 @@ declare const ollamaChatStreamChunkSchema: z.ZodDiscriminatedUnion<"done", [z.Zo | |
| 5799 5804 | 
             
                    content: string;
         | 
| 5800 5805 | 
             
                }>;
         | 
| 5801 5806 | 
             
            }, "strip", z.ZodTypeAny, {
         | 
| 5807 | 
            +
                model: string;
         | 
| 5802 5808 | 
             
                message: {
         | 
| 5803 5809 | 
             
                    role: string;
         | 
| 5804 5810 | 
             
                    content: string;
         | 
| 5805 5811 | 
             
                };
         | 
| 5806 | 
            -
                model: string;
         | 
| 5807 5812 | 
             
                done: false;
         | 
| 5808 5813 | 
             
                created_at: string;
         | 
| 5809 5814 | 
             
            }, {
         | 
| 5815 | 
            +
                model: string;
         | 
| 5810 5816 | 
             
                message: {
         | 
| 5811 5817 | 
             
                    role: string;
         | 
| 5812 5818 | 
             
                    content: string;
         | 
| 5813 5819 | 
             
                };
         | 
| 5814 | 
            -
                model: string;
         | 
| 5815 5820 | 
             
                done: false;
         | 
| 5816 5821 | 
             
                created_at: string;
         | 
| 5817 5822 | 
             
            }>, z.ZodObject<{
         | 
| @@ -5861,11 +5866,11 @@ declare const OllamaChatResponseFormat: { | |
| 5861 5866 | 
             
                        requestBodyValues: unknown;
         | 
| 5862 5867 | 
             
                        response: Response;
         | 
| 5863 5868 | 
             
                    }) => Promise<{
         | 
| 5869 | 
            +
                        model: string;
         | 
| 5864 5870 | 
             
                        message: {
         | 
| 5865 5871 | 
             
                            role: string;
         | 
| 5866 5872 | 
             
                            content: string;
         | 
| 5867 5873 | 
             
                        };
         | 
| 5868 | 
            -
                        model: string;
         | 
| 5869 5874 | 
             
                        done: true;
         | 
| 5870 5875 | 
             
                        created_at: string;
         | 
| 5871 5876 | 
             
                        total_duration: number;
         | 
| @@ -5885,11 +5890,11 @@ declare const OllamaChatResponseFormat: { | |
| 5885 5890 | 
             
                    handler: ({ response }: {
         | 
| 5886 5891 | 
             
                        response: Response;
         | 
| 5887 5892 | 
             
                    }) => Promise<AsyncIterable<Delta<{
         | 
| 5893 | 
            +
                        model: string;
         | 
| 5888 5894 | 
             
                        message: {
         | 
| 5889 5895 | 
             
                            role: string;
         | 
| 5890 5896 | 
             
                            content: string;
         | 
| 5891 5897 | 
             
                        };
         | 
| 5892 | 
            -
                        model: string;
         | 
| 5893 5898 | 
             
                        done: false;
         | 
| 5894 5899 | 
             
                        created_at: string;
         | 
| 5895 5900 | 
             
                    } | {
         | 
| @@ -7424,12 +7429,12 @@ declare abstract class AbstractOpenAITextEmbeddingModel<SETTINGS extends Abstrac | |
| 7424 7429 | 
             
                doEmbedValues(texts: string[], callOptions: FunctionCallOptions): Promise<{
         | 
| 7425 7430 | 
             
                    rawResponse: {
         | 
| 7426 7431 | 
             
                        object: "list";
         | 
| 7432 | 
            +
                        model: string;
         | 
| 7427 7433 | 
             
                        data: {
         | 
| 7428 7434 | 
             
                            object: "embedding";
         | 
| 7429 7435 | 
             
                            embedding: number[];
         | 
| 7430 7436 | 
             
                            index: number;
         | 
| 7431 7437 | 
             
                        }[];
         | 
| 7432 | 
            -
                        model: string;
         | 
| 7433 7438 | 
             
                        usage?: {
         | 
| 7434 7439 | 
             
                            prompt_tokens: number;
         | 
| 7435 7440 | 
             
                            total_tokens: number;
         | 
| @@ -7438,7 +7443,7 @@ declare abstract class AbstractOpenAITextEmbeddingModel<SETTINGS extends Abstrac | |
| 7438 7443 | 
             
                    embeddings: number[][];
         | 
| 7439 7444 | 
             
                }>;
         | 
| 7440 7445 | 
             
            }
         | 
| 7441 | 
            -
            declare const openAITextEmbeddingResponseSchema: z.ZodObject<{
         | 
| 7446 | 
            +
            declare const openAITextEmbeddingResponseSchema$1: z.ZodObject<{
         | 
| 7442 7447 | 
             
                object: z.ZodLiteral<"list">;
         | 
| 7443 7448 | 
             
                data: z.ZodArray<z.ZodObject<{
         | 
| 7444 7449 | 
             
                    object: z.ZodLiteral<"embedding">;
         | 
| @@ -7466,30 +7471,30 @@ declare const openAITextEmbeddingResponseSchema: z.ZodObject<{ | |
| 7466 7471 | 
             
                }>>;
         | 
| 7467 7472 | 
             
            }, "strip", z.ZodTypeAny, {
         | 
| 7468 7473 | 
             
                object: "list";
         | 
| 7474 | 
            +
                model: string;
         | 
| 7469 7475 | 
             
                data: {
         | 
| 7470 7476 | 
             
                    object: "embedding";
         | 
| 7471 7477 | 
             
                    embedding: number[];
         | 
| 7472 7478 | 
             
                    index: number;
         | 
| 7473 7479 | 
             
                }[];
         | 
| 7474 | 
            -
                model: string;
         | 
| 7475 7480 | 
             
                usage?: {
         | 
| 7476 7481 | 
             
                    prompt_tokens: number;
         | 
| 7477 7482 | 
             
                    total_tokens: number;
         | 
| 7478 7483 | 
             
                } | undefined;
         | 
| 7479 7484 | 
             
            }, {
         | 
| 7480 7485 | 
             
                object: "list";
         | 
| 7486 | 
            +
                model: string;
         | 
| 7481 7487 | 
             
                data: {
         | 
| 7482 7488 | 
             
                    object: "embedding";
         | 
| 7483 7489 | 
             
                    embedding: number[];
         | 
| 7484 7490 | 
             
                    index: number;
         | 
| 7485 7491 | 
             
                }[];
         | 
| 7486 | 
            -
                model: string;
         | 
| 7487 7492 | 
             
                usage?: {
         | 
| 7488 7493 | 
             
                    prompt_tokens: number;
         | 
| 7489 7494 | 
             
                    total_tokens: number;
         | 
| 7490 7495 | 
             
                } | undefined;
         | 
| 7491 7496 | 
             
            }>;
         | 
| 7492 | 
            -
            type OpenAITextEmbeddingResponse = z.infer<typeof openAITextEmbeddingResponseSchema>;
         | 
| 7497 | 
            +
            type OpenAITextEmbeddingResponse = z.infer<typeof openAITextEmbeddingResponseSchema$1>;
         | 
| 7493 7498 |  | 
| 7494 7499 | 
             
            type AzureOpenAIApiConfigurationOptions = {
         | 
| 7495 7500 | 
             
                resourceName: string;
         | 
| @@ -7728,21 +7733,12 @@ OpenAIChatSettings> { | |
| 7728 7733 | 
             
            declare const OPENAI_TEXT_GENERATION_MODELS: {
         | 
| 7729 7734 | 
             
                "gpt-3.5-turbo-instruct": {
         | 
| 7730 7735 | 
             
                    contextWindowSize: number;
         | 
| 7731 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7732 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7733 7736 | 
             
                };
         | 
| 7734 7737 | 
             
            };
         | 
| 7735 7738 | 
             
            declare function getOpenAICompletionModelInformation(model: OpenAICompletionModelType): {
         | 
| 7736 7739 | 
             
                contextWindowSize: number;
         | 
| 7737 | 
            -
                promptTokenCostInMillicents: number;
         | 
| 7738 | 
            -
                completionTokenCostInMillicents: number;
         | 
| 7739 7740 | 
             
            };
         | 
| 7740 7741 | 
             
            type OpenAICompletionModelType = keyof typeof OPENAI_TEXT_GENERATION_MODELS;
         | 
| 7741 | 
            -
            declare const isOpenAICompletionModel: (model: string) => model is "gpt-3.5-turbo-instruct";
         | 
| 7742 | 
            -
            declare const calculateOpenAICompletionCostInMillicents: ({ model, response, }: {
         | 
| 7743 | 
            -
                model: OpenAICompletionModelType;
         | 
| 7744 | 
            -
                response: OpenAICompletionResponse;
         | 
| 7745 | 
            -
            }) => number;
         | 
| 7746 7742 | 
             
            interface OpenAICompletionModelSettings extends AbstractOpenAICompletionModelSettings {
         | 
| 7747 7743 | 
             
                model: OpenAICompletionModelType;
         | 
| 7748 7744 | 
             
            }
         | 
| @@ -7786,28 +7782,71 @@ declare const OPENAI_TEXT_EMBEDDING_MODELS: { | |
| 7786 7782 | 
             
                "text-embedding-3-small": {
         | 
| 7787 7783 | 
             
                    contextWindowSize: number;
         | 
| 7788 7784 | 
             
                    dimensions: number;
         | 
| 7789 | 
            -
                    tokenCostInMillicents: number;
         | 
| 7790 7785 | 
             
                };
         | 
| 7791 7786 | 
             
                "text-embedding-3-large": {
         | 
| 7792 7787 | 
             
                    contextWindowSize: number;
         | 
| 7793 7788 | 
             
                    dimensions: number;
         | 
| 7794 | 
            -
                    tokenCostInMillicents: number;
         | 
| 7795 7789 | 
             
                };
         | 
| 7796 7790 | 
             
                "text-embedding-ada-002": {
         | 
| 7797 7791 | 
             
                    contextWindowSize: number;
         | 
| 7798 7792 | 
             
                    dimensions: number;
         | 
| 7799 | 
            -
                    tokenCostInMillicents: number;
         | 
| 7800 7793 | 
             
                };
         | 
| 7801 7794 | 
             
            };
         | 
| 7802 7795 | 
             
            type OpenAITextEmbeddingModelType = keyof typeof OPENAI_TEXT_EMBEDDING_MODELS;
         | 
| 7803 | 
            -
            declare const isOpenAIEmbeddingModel: (model: string) => model is "text-embedding-3-small" | "text-embedding-3-large" | "text-embedding-ada-002";
         | 
| 7804 | 
            -
            declare const calculateOpenAIEmbeddingCostInMillicents: ({ model, responses, }: {
         | 
| 7805 | 
            -
                model: OpenAITextEmbeddingModelType;
         | 
| 7806 | 
            -
                responses: OpenAITextEmbeddingResponse[];
         | 
| 7807 | 
            -
            }) => number;
         | 
| 7808 7796 | 
             
            interface OpenAITextEmbeddingModelSettings extends AbstractOpenAITextEmbeddingModelSettings {
         | 
| 7809 7797 | 
             
                model: OpenAITextEmbeddingModelType;
         | 
| 7810 7798 | 
             
            }
         | 
| 7799 | 
            +
            declare const openAITextEmbeddingResponseSchema: z$1.ZodObject<{
         | 
| 7800 | 
            +
                object: z$1.ZodLiteral<"list">;
         | 
| 7801 | 
            +
                data: z$1.ZodArray<z$1.ZodObject<{
         | 
| 7802 | 
            +
                    object: z$1.ZodLiteral<"embedding">;
         | 
| 7803 | 
            +
                    embedding: z$1.ZodArray<z$1.ZodNumber, "many">;
         | 
| 7804 | 
            +
                    index: z$1.ZodNumber;
         | 
| 7805 | 
            +
                }, "strip", z$1.ZodTypeAny, {
         | 
| 7806 | 
            +
                    object: "embedding";
         | 
| 7807 | 
            +
                    embedding: number[];
         | 
| 7808 | 
            +
                    index: number;
         | 
| 7809 | 
            +
                }, {
         | 
| 7810 | 
            +
                    object: "embedding";
         | 
| 7811 | 
            +
                    embedding: number[];
         | 
| 7812 | 
            +
                    index: number;
         | 
| 7813 | 
            +
                }>, "many">;
         | 
| 7814 | 
            +
                model: z$1.ZodString;
         | 
| 7815 | 
            +
                usage: z$1.ZodOptional<z$1.ZodObject<{
         | 
| 7816 | 
            +
                    prompt_tokens: z$1.ZodNumber;
         | 
| 7817 | 
            +
                    total_tokens: z$1.ZodNumber;
         | 
| 7818 | 
            +
                }, "strip", z$1.ZodTypeAny, {
         | 
| 7819 | 
            +
                    prompt_tokens: number;
         | 
| 7820 | 
            +
                    total_tokens: number;
         | 
| 7821 | 
            +
                }, {
         | 
| 7822 | 
            +
                    prompt_tokens: number;
         | 
| 7823 | 
            +
                    total_tokens: number;
         | 
| 7824 | 
            +
                }>>;
         | 
| 7825 | 
            +
            }, "strip", z$1.ZodTypeAny, {
         | 
| 7826 | 
            +
                object: "list";
         | 
| 7827 | 
            +
                model: string;
         | 
| 7828 | 
            +
                data: {
         | 
| 7829 | 
            +
                    object: "embedding";
         | 
| 7830 | 
            +
                    embedding: number[];
         | 
| 7831 | 
            +
                    index: number;
         | 
| 7832 | 
            +
                }[];
         | 
| 7833 | 
            +
                usage?: {
         | 
| 7834 | 
            +
                    prompt_tokens: number;
         | 
| 7835 | 
            +
                    total_tokens: number;
         | 
| 7836 | 
            +
                } | undefined;
         | 
| 7837 | 
            +
            }, {
         | 
| 7838 | 
            +
                object: "list";
         | 
| 7839 | 
            +
                model: string;
         | 
| 7840 | 
            +
                data: {
         | 
| 7841 | 
            +
                    object: "embedding";
         | 
| 7842 | 
            +
                    embedding: number[];
         | 
| 7843 | 
            +
                    index: number;
         | 
| 7844 | 
            +
                }[];
         | 
| 7845 | 
            +
                usage?: {
         | 
| 7846 | 
            +
                    prompt_tokens: number;
         | 
| 7847 | 
            +
                    total_tokens: number;
         | 
| 7848 | 
            +
                } | undefined;
         | 
| 7849 | 
            +
            }>;
         | 
| 7811 7850 | 
             
            /**
         | 
| 7812 7851 | 
             
             * Create a text embedding model that calls the OpenAI embedding API.
         | 
| 7813 7852 | 
             
             *
         | 
| @@ -7866,115 +7905,34 @@ declare class TikTokenTokenizer implements FullTokenizer { | |
| 7866 7905 | 
             
                detokenize(tokens: number[]): Promise<string>;
         | 
| 7867 7906 | 
             
            }
         | 
| 7868 7907 |  | 
| 7869 | 
            -
            declare const  | 
| 7870 | 
            -
                "gpt-4":  | 
| 7871 | 
            -
             | 
| 7872 | 
            -
             | 
| 7873 | 
            -
             | 
| 7874 | 
            -
                 | 
| 7875 | 
            -
                "gpt-4- | 
| 7876 | 
            -
             | 
| 7877 | 
            -
             | 
| 7878 | 
            -
             | 
| 7879 | 
            -
                 | 
| 7880 | 
            -
                "gpt- | 
| 7881 | 
            -
             | 
| 7882 | 
            -
             | 
| 7883 | 
            -
             | 
| 7884 | 
            -
             | 
| 7885 | 
            -
             | 
| 7886 | 
            -
                 | 
| 7887 | 
            -
                "gpt-4-turbo-preview": {
         | 
| 7888 | 
            -
                    contextWindowSize: number;
         | 
| 7889 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7890 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7891 | 
            -
                };
         | 
| 7892 | 
            -
                "gpt-4-1106-preview": {
         | 
| 7893 | 
            -
                    contextWindowSize: number;
         | 
| 7894 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7895 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7896 | 
            -
                };
         | 
| 7897 | 
            -
                "gpt-4-0125-preview": {
         | 
| 7898 | 
            -
                    contextWindowSize: number;
         | 
| 7899 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7900 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7901 | 
            -
                };
         | 
| 7902 | 
            -
                "gpt-4-vision-preview": {
         | 
| 7903 | 
            -
                    contextWindowSize: number;
         | 
| 7904 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7905 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7906 | 
            -
                };
         | 
| 7907 | 
            -
                "gpt-4-32k": {
         | 
| 7908 | 
            -
                    contextWindowSize: number;
         | 
| 7909 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7910 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7911 | 
            -
                };
         | 
| 7912 | 
            -
                "gpt-4-32k-0314": {
         | 
| 7913 | 
            -
                    contextWindowSize: number;
         | 
| 7914 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7915 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7916 | 
            -
                };
         | 
| 7917 | 
            -
                "gpt-4-32k-0613": {
         | 
| 7918 | 
            -
                    contextWindowSize: number;
         | 
| 7919 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7920 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7921 | 
            -
                };
         | 
| 7922 | 
            -
                "gpt-3.5-turbo": {
         | 
| 7923 | 
            -
                    contextWindowSize: number;
         | 
| 7924 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7925 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7926 | 
            -
                    fineTunedPromptTokenCostInMillicents: number;
         | 
| 7927 | 
            -
                    fineTunedCompletionTokenCostInMillicents: number;
         | 
| 7928 | 
            -
                };
         | 
| 7929 | 
            -
                "gpt-3.5-turbo-0125": {
         | 
| 7930 | 
            -
                    contextWindowSize: number;
         | 
| 7931 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7932 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7933 | 
            -
                };
         | 
| 7934 | 
            -
                "gpt-3.5-turbo-1106": {
         | 
| 7935 | 
            -
                    contextWindowSize: number;
         | 
| 7936 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7937 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7938 | 
            -
                };
         | 
| 7939 | 
            -
                "gpt-3.5-turbo-0301": {
         | 
| 7940 | 
            -
                    contextWindowSize: number;
         | 
| 7941 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7942 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7943 | 
            -
                };
         | 
| 7944 | 
            -
                "gpt-3.5-turbo-0613": {
         | 
| 7945 | 
            -
                    contextWindowSize: number;
         | 
| 7946 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7947 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7948 | 
            -
                    fineTunedPromptTokenCostInMillicents: number;
         | 
| 7949 | 
            -
                    fineTunedCompletionTokenCostInMillicents: number;
         | 
| 7950 | 
            -
                };
         | 
| 7951 | 
            -
                "gpt-3.5-turbo-16k": {
         | 
| 7952 | 
            -
                    contextWindowSize: number;
         | 
| 7953 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7954 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7955 | 
            -
                };
         | 
| 7956 | 
            -
                "gpt-3.5-turbo-16k-0613": {
         | 
| 7957 | 
            -
                    contextWindowSize: number;
         | 
| 7958 | 
            -
                    promptTokenCostInMillicents: number;
         | 
| 7959 | 
            -
                    completionTokenCostInMillicents: number;
         | 
| 7960 | 
            -
                };
         | 
| 7908 | 
            +
            declare const CHAT_MODEL_CONTEXT_WINDOW_SIZES: {
         | 
| 7909 | 
            +
                readonly "gpt-4": 8192;
         | 
| 7910 | 
            +
                readonly "gpt-4-0314": 8192;
         | 
| 7911 | 
            +
                readonly "gpt-4-0613": 8192;
         | 
| 7912 | 
            +
                readonly "gpt-4-turbo-preview": 128000;
         | 
| 7913 | 
            +
                readonly "gpt-4-1106-preview": 128000;
         | 
| 7914 | 
            +
                readonly "gpt-4-0125-preview": 128000;
         | 
| 7915 | 
            +
                readonly "gpt-4-vision-preview": 128000;
         | 
| 7916 | 
            +
                readonly "gpt-4-32k": 32768;
         | 
| 7917 | 
            +
                readonly "gpt-4-32k-0314": 32768;
         | 
| 7918 | 
            +
                readonly "gpt-4-32k-0613": 32768;
         | 
| 7919 | 
            +
                readonly "gpt-3.5-turbo": 4096;
         | 
| 7920 | 
            +
                readonly "gpt-3.5-turbo-0125": 16385;
         | 
| 7921 | 
            +
                readonly "gpt-3.5-turbo-1106": 16385;
         | 
| 7922 | 
            +
                readonly "gpt-3.5-turbo-0301": 4096;
         | 
| 7923 | 
            +
                readonly "gpt-3.5-turbo-0613": 4096;
         | 
| 7924 | 
            +
                readonly "gpt-3.5-turbo-16k": 16384;
         | 
| 7925 | 
            +
                readonly "gpt-3.5-turbo-16k-0613": 16384;
         | 
| 7961 7926 | 
             
            };
         | 
| 7962 7927 | 
             
            declare function getOpenAIChatModelInformation(model: OpenAIChatModelType): {
         | 
| 7963 7928 | 
             
                baseModel: OpenAIChatBaseModelType;
         | 
| 7964 7929 | 
             
                isFineTuned: boolean;
         | 
| 7965 7930 | 
             
                contextWindowSize: number;
         | 
| 7966 | 
            -
                promptTokenCostInMillicents: number | null;
         | 
| 7967 | 
            -
                completionTokenCostInMillicents: number | null;
         | 
| 7968 7931 | 
             
            };
         | 
| 7969 7932 | 
             
            type FineTuneableOpenAIChatModelType = `gpt-3.5-turbo` | `gpt-3.5-turbo-0613` | `gpt-4-0613`;
         | 
| 7970 7933 | 
             
            type FineTunedOpenAIChatModelType = `ft:${FineTuneableOpenAIChatModelType}:${string}:${string}:${string}`;
         | 
| 7971 | 
            -
            type OpenAIChatBaseModelType = keyof typeof  | 
| 7934 | 
            +
            type OpenAIChatBaseModelType = keyof typeof CHAT_MODEL_CONTEXT_WINDOW_SIZES;
         | 
| 7972 7935 | 
             
            type OpenAIChatModelType = OpenAIChatBaseModelType | FineTunedOpenAIChatModelType;
         | 
| 7973 | 
            -
            declare const isOpenAIChatModel: (model: string) => model is OpenAIChatModelType;
         | 
| 7974 | 
            -
            declare const calculateOpenAIChatCostInMillicents: ({ model, response, }: {
         | 
| 7975 | 
            -
                model: OpenAIChatModelType;
         | 
| 7976 | 
            -
                response: OpenAIChatResponse;
         | 
| 7977 | 
            -
            }) => number | null;
         | 
| 7978 7936 | 
             
            interface OpenAIChatSettings extends AbstractOpenAIChatSettings {
         | 
| 7979 7937 | 
             
                model: OpenAIChatModelType;
         | 
| 7980 7938 | 
             
            }
         | 
| @@ -8056,24 +8014,8 @@ declare const openAIErrorDataSchema: z.ZodObject<{ | |
| 8056 8014 | 
             
            }>;
         | 
| 8057 8015 | 
             
            type OpenAIErrorData = z.infer<typeof openAIErrorDataSchema>;
         | 
| 8058 8016 |  | 
| 8059 | 
            -
            declare const OPENAI_IMAGE_MODELS: {
         | 
| 8060 | 
            -
                "dall-e-2": {
         | 
| 8061 | 
            -
                    getCost(settings: OpenAIImageGenerationSettings): 2000 | 1800 | 1600 | null;
         | 
| 8062 | 
            -
                };
         | 
| 8063 | 
            -
                "dall-e-3": {
         | 
| 8064 | 
            -
                    getCost(settings: OpenAIImageGenerationSettings): 4000 | 8000 | 12000 | null;
         | 
| 8065 | 
            -
                };
         | 
| 8066 | 
            -
            };
         | 
| 8067 | 
            -
            /**
         | 
| 8068 | 
            -
             * @see https://openai.com/pricing
         | 
| 8069 | 
            -
             */
         | 
| 8070 | 
            -
            declare const calculateOpenAIImageGenerationCostInMillicents: ({ model, settings, }: {
         | 
| 8071 | 
            -
                model: OpenAIImageModelType;
         | 
| 8072 | 
            -
                settings: OpenAIImageGenerationSettings;
         | 
| 8073 | 
            -
            }) => number | null;
         | 
| 8074 | 
            -
            type OpenAIImageModelType = keyof typeof OPENAI_IMAGE_MODELS;
         | 
| 8075 8017 | 
             
            interface OpenAIImageGenerationCallSettings {
         | 
| 8076 | 
            -
                model:  | 
| 8018 | 
            +
                model: "dall-e-2" | "dall-e-3";
         | 
| 8077 8019 | 
             
                size?: "256x256" | "512x512" | "1024x1024" | "1792x1024" | "1024x1792";
         | 
| 8078 8020 | 
             
                quality?: "standard" | "hd";
         | 
| 8079 8021 | 
             
                style?: "vivid" | "natural";
         | 
| @@ -8180,24 +8122,9 @@ declare const OpenAIImageGenerationResponseFormat: { | |
| 8180 8122 | 
             
                };
         | 
| 8181 8123 | 
             
            };
         | 
| 8182 8124 |  | 
| 8183 | 
            -
            /**
         | 
| 8184 | 
            -
             * @see https://openai.com/pricing
         | 
| 8185 | 
            -
             */
         | 
| 8186 | 
            -
            declare const OPENAI_SPEECH_MODELS: {
         | 
| 8187 | 
            -
                "tts-1": {
         | 
| 8188 | 
            -
                    costInMillicentsPerCharacter: number;
         | 
| 8189 | 
            -
                };
         | 
| 8190 | 
            -
                "tts-1-hd": {
         | 
| 8191 | 
            -
                    costInMillicentsPerCharacter: number;
         | 
| 8192 | 
            -
                };
         | 
| 8193 | 
            -
            };
         | 
| 8194 | 
            -
            type OpenAISpeechModelType = keyof typeof OPENAI_SPEECH_MODELS;
         | 
| 8195 | 
            -
            declare const calculateOpenAISpeechCostInMillicents: ({ model, input, }: {
         | 
| 8196 | 
            -
                model: OpenAISpeechModelType;
         | 
| 8197 | 
            -
                input: string;
         | 
| 8198 | 
            -
            }) => number | null;
         | 
| 8199 8125 | 
             
            type OpenAISpeechVoice = "alloy" | "echo" | "fable" | "onyx" | "nova" | "shimmer";
         | 
| 8200 8126 | 
             
            type OpenAISpeechModelResponseFormat = "mp3" | "opus" | "aac" | "flac";
         | 
| 8127 | 
            +
            type OpenAISpeechModelType = "tts-1" | "tts-1-hd";
         | 
| 8201 8128 | 
             
            interface OpenAISpeechModelSettings extends SpeechGenerationModelSettings {
         | 
| 8202 8129 | 
             
                api?: ApiConfiguration;
         | 
| 8203 8130 | 
             
                voice: OpenAISpeechVoice;
         | 
| @@ -8220,26 +8147,14 @@ declare class OpenAISpeechModel extends AbstractModel<OpenAISpeechModelSettings> | |
| 8220 8147 | 
             
                constructor(settings: OpenAISpeechModelSettings);
         | 
| 8221 8148 | 
             
                readonly provider: "openai";
         | 
| 8222 8149 | 
             
                get voice(): OpenAISpeechVoice;
         | 
| 8223 | 
            -
                get modelName():  | 
| 8150 | 
            +
                get modelName(): OpenAISpeechModelType;
         | 
| 8224 8151 | 
             
                private callAPI;
         | 
| 8225 8152 | 
             
                get settingsForEvent(): Partial<OpenAISpeechModelSettings>;
         | 
| 8226 8153 | 
             
                doGenerateSpeechStandard(text: string, options: FunctionCallOptions): Promise<Uint8Array>;
         | 
| 8227 8154 | 
             
                withSettings(additionalSettings: Partial<OpenAISpeechModelSettings>): this;
         | 
| 8228 8155 | 
             
            }
         | 
| 8229 8156 |  | 
| 8230 | 
            -
             | 
| 8231 | 
            -
             * @see https://openai.com/pricing
         | 
| 8232 | 
            -
             */
         | 
| 8233 | 
            -
            declare const OPENAI_TRANSCRIPTION_MODELS: {
         | 
| 8234 | 
            -
                "whisper-1": {
         | 
| 8235 | 
            -
                    costInMillicentsPerSecond: number;
         | 
| 8236 | 
            -
                };
         | 
| 8237 | 
            -
            };
         | 
| 8238 | 
            -
            type OpenAITranscriptionModelType = keyof typeof OPENAI_TRANSCRIPTION_MODELS;
         | 
| 8239 | 
            -
            declare const calculateOpenAITranscriptionCostInMillicents: ({ model, response, }: {
         | 
| 8240 | 
            -
                model: OpenAITranscriptionModelType;
         | 
| 8241 | 
            -
                response: OpenAITranscriptionVerboseJsonResponse;
         | 
| 8242 | 
            -
            }) => number | null;
         | 
| 8157 | 
            +
            type OpenAITranscriptionModelType = "whisper-1";
         | 
| 8243 8158 | 
             
            interface OpenAITranscriptionModelSettings extends TranscriptionModelSettings {
         | 
| 8244 8159 | 
             
                api?: ApiConfiguration;
         | 
| 8245 8160 | 
             
                /**
         | 
| @@ -8290,8 +8205,8 @@ declare class OpenAITranscriptionModel extends AbstractModel<OpenAITranscription | |
| 8290 8205 | 
             
                    rawResponse: {
         | 
| 8291 8206 | 
             
                        text: string;
         | 
| 8292 8207 | 
             
                        duration: number;
         | 
| 8293 | 
            -
                        task: "transcribe";
         | 
| 8294 8208 | 
             
                        language: string;
         | 
| 8209 | 
            +
                        task: "transcribe";
         | 
| 8295 8210 | 
             
                        segments: {
         | 
| 8296 8211 | 
             
                            text: string;
         | 
| 8297 8212 | 
             
                            id: number;
         | 
| @@ -8370,8 +8285,8 @@ declare const openAITranscriptionVerboseJsonSchema: z.ZodObject<{ | |
| 8370 8285 | 
             
            }, "strip", z.ZodTypeAny, {
         | 
| 8371 8286 | 
             
                text: string;
         | 
| 8372 8287 | 
             
                duration: number;
         | 
| 8373 | 
            -
                task: "transcribe";
         | 
| 8374 8288 | 
             
                language: string;
         | 
| 8289 | 
            +
                task: "transcribe";
         | 
| 8375 8290 | 
             
                segments: {
         | 
| 8376 8291 | 
             
                    text: string;
         | 
| 8377 8292 | 
             
                    id: number;
         | 
| @@ -8388,8 +8303,8 @@ declare const openAITranscriptionVerboseJsonSchema: z.ZodObject<{ | |
| 8388 8303 | 
             
            }, {
         | 
| 8389 8304 | 
             
                text: string;
         | 
| 8390 8305 | 
             
                duration: number;
         | 
| 8391 | 
            -
                task: "transcribe";
         | 
| 8392 8306 | 
             
                language: string;
         | 
| 8307 | 
            +
                task: "transcribe";
         | 
| 8393 8308 | 
             
                segments: {
         | 
| 8394 8309 | 
             
                    text: string;
         | 
| 8395 8310 | 
             
                    id: number;
         | 
| @@ -8421,8 +8336,8 @@ declare const OpenAITranscriptionResponseFormat: { | |
| 8421 8336 | 
             
                    handler: ResponseHandler<{
         | 
| 8422 8337 | 
             
                        text: string;
         | 
| 8423 8338 | 
             
                        duration: number;
         | 
| 8424 | 
            -
                        task: "transcribe";
         | 
| 8425 8339 | 
             
                        language: string;
         | 
| 8340 | 
            +
                        task: "transcribe";
         | 
| 8426 8341 | 
             
                        segments: {
         | 
| 8427 8342 | 
             
                            text: string;
         | 
| 8428 8343 | 
             
                            id: number;
         | 
| @@ -9544,4 +9459,4 @@ declare function upsertIntoVectorIndex<VALUE, OBJECT>({ vectorIndex, embeddingMo | |
| 9544 9459 | 
             
                getId?: (object: OBJECT, index: number) => string | undefined;
         | 
| 9545 9460 | 
             
            }, options?: FunctionOptions): Promise<void>;
         | 
| 9546 9461 |  | 
| 9547 | 
            -
            export { AbortError, AbstractOpenAIChatModel, type AbstractOpenAIChatSettings, AbstractOpenAICompletionModel, type AbstractOpenAICompletionModelSettings, AbstractOpenAITextEmbeddingModel, type AbstractOpenAITextEmbeddingModelSettings, AlpacaPromptTemplate as AlpacaPrompt, ApiCallError, type ApiConfiguration, type AssistantContent, AsyncQueue, type AudioMimeType, Automatic1111ApiConfiguration, type Automatic1111ErrorData, Automatic1111ImageGenerationModel, type Automatic1111ImageGenerationPrompt, type Automatic1111ImageGenerationResponse, type Automatic1111ImageGenerationSettings, AzureOpenAIApiConfiguration, type AzureOpenAIApiConfigurationOptions, type BaseFunctionEvent, type BaseFunctionFinishedEvent, type BaseFunctionFinishedEventResult, type BaseFunctionStartedEvent, type BaseModelCallFinishedEvent, type BaseModelCallFinishedEventResult, type BaseModelCallStartedEvent, BaseUrlApiConfiguration, BaseUrlApiConfigurationWithDefaults, type BaseUrlPartsApiConfigurationOptions, type BasicTokenizer, COHERE_TEXT_EMBEDDING_MODELS, COHERE_TEXT_GENERATION_MODELS, type Cache, ChatMLPromptTemplate as ChatMLPrompt, ChatMessage, type ChatPrompt, type Classifier, type ClassifierSettings, type ClassifyFinishedEvent, type ClassifyFinishedEventResult, type ClassifyStartedEvent, CohereApiConfiguration, type CohereDetokenizationResponse, type CohereErrorData, CohereTextEmbeddingModel, type CohereTextEmbeddingModelSettings, type CohereTextEmbeddingModelType, type CohereTextEmbeddingResponse, CohereTextGenerationModel, type CohereTextGenerationModelSettings, type CohereTextGenerationModelType, type CohereTextGenerationResponse, CohereTextGenerationResponseFormat, type CohereTextGenerationResponseFormatType, type CohereTextStreamChunk, type CohereTokenizationResponse, CohereTokenizer, type CohereTokenizerModelType, type CohereTokenizerSettings, type CustomHeaderProvider, type DataContent, DefaultRun, type Delta, ElevenLabsApiConfiguration, ElevenLabsSpeechModel, type ElevenLabsSpeechModelSettings, type EmbeddingFinishedEvent, type EmbeddingFinishedEventResult, type EmbeddingModel, type EmbeddingModelSettings, EmbeddingSimilarityClassifier, type EmbeddingSimilarityClassifierSettings, type EmbeddingStartedEvent, type ExecuteFunctionFinishedEvent, type ExecuteFunctionStartedEvent, type ExecuteToolFinishedEvent, type ExecuteToolMetadata, type ExecuteToolStartedEvent, FireworksAIApiConfiguration, type FlexibleObjectFromTextPromptTemplate, type FullTokenizer, type FunctionCallOptions, type FunctionEvent, FunctionEventSource, type FunctionObserver, type FunctionOptions, type HasContextWindowSize, type HasTokenizer, type HeaderParameters, HeliconeOpenAIApiConfiguration, HuggingFaceApiConfiguration, type HuggingFaceErrorData, HuggingFaceTextEmbeddingModel, type HuggingFaceTextEmbeddingModelSettings, type HuggingFaceTextEmbeddingResponse, HuggingFaceTextGenerationModel, type HuggingFaceTextGenerationModelSettings, type HuggingFaceTextGenerationResponse, type ImageGenerationFinishedEvent, type ImageGenerationFinishedEventResult, type ImageGenerationModel, type ImageGenerationModelSettings, type ImageGenerationStartedEvent, type ImagePart, type InstructionContent, type InstructionPrompt, InvalidPromptError, JSONParseError, type JsonSchemaProducer, Llama2PromptTemplate as Llama2Prompt, LlamaCppApiConfiguration, LlamaCppCompletionModel, type LlamaCppCompletionModelSettings, type LlamaCppCompletionPrompt, LlamaCppCompletionResponseFormat, type LlamaCppCompletionResponseFormatType, type LlamaCppErrorData, LlamaCppTextEmbeddingModel, type LlamaCppTextEmbeddingModelSettings, type LlamaCppTextEmbeddingResponse, type LlamaCppTextGenerationResponse, type LlamaCppTextStreamChunk, type LlamaCppTokenizationResponse, LlamaCppTokenizer, LmntApiConfiguration, LmntSpeechModel, type LmntSpeechModelSettings, type LmntSpeechResponse, type LogFormat, MemoryCache, MemoryVectorIndex, MistralApiConfiguration, type MistralChatMessage, MistralChatModel, type MistralChatModelSettings, type MistralChatPrompt, type MistralChatResponse, MistralChatResponseFormat, type MistralChatResponseFormatType, type MistralChatStreamChunk, type MistralErrorData, MistralInstructPromptTemplate as MistralInstructPrompt, MistralTextEmbeddingModel, type MistralTextEmbeddingModelSettings, type MistralTextEmbeddingResponse, type Model, type ModelCallFinishedEvent, type ModelCallMetadata, type ModelCallStartedEvent, type ModelInformation, type ModelSettings, NeuralChatPromptTemplate as NeuralChatPrompt, NoSuchToolDefinitionError, OPENAI_CHAT_MESSAGE_BASE_TOKEN_COUNT, OPENAI_CHAT_MODELS, OPENAI_CHAT_PROMPT_BASE_TOKEN_COUNT, OPENAI_IMAGE_MODELS, OPENAI_SPEECH_MODELS, OPENAI_TEXT_EMBEDDING_MODELS, OPENAI_TEXT_GENERATION_MODELS, OPENAI_TRANSCRIPTION_MODELS, ObjectFromTextGenerationModel, type ObjectFromTextPromptTemplate, ObjectFromTextStreamingModel, type ObjectGenerationFinishedEvent, type ObjectGenerationFinishedEventResult, type ObjectGenerationModel, type ObjectGenerationModelSettings, type ObjectGenerationStartedEvent, ObjectGeneratorTool, ObjectParseError, type ObjectStream, ObjectStreamFromResponse, ObjectStreamResponse, type ObjectStreamingFinishedEvent, type ObjectStreamingModel, type ObjectStreamingStartedEvent, ObjectValidationError, OllamaApiConfiguration, type OllamaChatMessage, OllamaChatModel, type OllamaChatModelSettings, type OllamaChatPrompt, type OllamaChatResponse, OllamaChatResponseFormat, type OllamaChatResponseFormatType, type OllamaChatStreamChunk, OllamaCompletionModel, type OllamaCompletionModelSettings, type OllamaCompletionPrompt$1 as OllamaCompletionPrompt, type OllamaCompletionResponse, OllamaCompletionResponseFormat, type OllamaCompletionResponseFormatType, type OllamaCompletionStreamChunk, type OllamaErrorData, OllamaTextEmbeddingModel, type OllamaTextEmbeddingModelSettings, type OllamaTextEmbeddingResponse, type OllamaTextGenerationSettings, OpenAIApiConfiguration, type OpenAIChatBaseModelType, type OpenAIChatChunk, OpenAIChatMessage, OpenAIChatModel, type OpenAIChatModelType, type OpenAIChatPrompt, type OpenAIChatResponse, OpenAIChatResponseFormat, type OpenAIChatResponseFormatType, type OpenAIChatSettings, type OpenAICompatibleApiConfiguration, OpenAICompatibleChatModel, type OpenAICompatibleChatSettings, OpenAICompatibleCompletionModel, type OpenAICompatibleCompletionModelSettings, type OpenAICompatibleProviderName, OpenAICompatibleTextEmbeddingModel, type OpenAICompatibleTextEmbeddingModelSettings, OpenAICompletionModel, type OpenAICompletionModelSettings, type OpenAICompletionModelType, type OpenAICompletionResponse, type OpenAIErrorData, type OpenAIImageGenerationBase64JsonResponse, type OpenAIImageGenerationCallSettings, OpenAIImageGenerationModel, OpenAIImageGenerationResponseFormat, type OpenAIImageGenerationResponseFormatType, type OpenAIImageGenerationSettings, type OpenAIImageGenerationUrlResponse, type OpenAIImageModelType, OpenAISpeechModel, type OpenAISpeechModelSettings, type OpenAISpeechModelType, type OpenAISpeechVoice, OpenAITextEmbeddingModel, type OpenAITextEmbeddingModelSettings, type OpenAITextEmbeddingModelType, type OpenAITextEmbeddingResponse, OpenAITextResponseFormat, type OpenAITextResponseFormatType, type OpenAITranscriptionJsonResponse, OpenAITranscriptionModel, type OpenAITranscriptionModelSettings, type OpenAITranscriptionModelType, OpenAITranscriptionResponseFormat, type OpenAITranscriptionResponseFormatType, type OpenAITranscriptionVerboseJsonResponse, type PartialBaseUrlPartsApiConfigurationOptions, PerplexityApiConfiguration, type PromptFunction, type PromptTemplate, PromptTemplateFullTextModel, PromptTemplateImageGenerationModel, PromptTemplateTextGenerationModel, PromptTemplateTextStreamingModel, type Retriever, RetryError, type RetryErrorReason, type RetryFunction, type Run, type Schema, type SpeechGenerationFinishedEvent, type SpeechGenerationFinishedEventResult, type SpeechGenerationModel, type SpeechGenerationModelSettings, type SpeechGenerationStartedEvent, type SpeechStreamingFinishedEvent, type SpeechStreamingStartedEvent, type SplitFunction, StabilityApiConfiguration, type StabilityClipGuidancePreset, type StabilityErrorData, StabilityImageGenerationModel, type StabilityImageGenerationModelType, type StabilityImageGenerationPrompt, type StabilityImageGenerationResponse, type StabilityImageGenerationSampler, type StabilityImageGenerationSettings, type StabilityImageGenerationStylePreset, type StreamingSpeechGenerationModel, SynthiaPromptTemplate as SynthiaPrompt, type TextChunk, type TextGenerationBaseModel, type TextGenerationFinishReason, type TextGenerationFinishedEvent, type TextGenerationFinishedEventResult, type TextGenerationModel, type TextGenerationModelSettings, type TextGenerationPromptTemplate, type TextGenerationPromptTemplateProvider, type TextGenerationResult, type TextGenerationStartedEvent, TextGenerationToolCallModel, TextGenerationToolCallsModel, type TextPart, TextPromptTemplate as TextPrompt, type TextStreamingBaseModel, type TextStreamingFinishedEvent, type TextStreamingModel, type TextStreamingStartedEvent, type ThrottleFunction, TikTokenTokenizer, type TikTokenTokenizerSettings, TogetherAIApiConfiguration, Tool, type ToolCall, ToolCallArgumentsValidationError, ToolCallError, ToolCallGenerationError, type ToolCallGenerationFinishedEvent, type ToolCallGenerationFinishedEventResult, type ToolCallGenerationModel, type ToolCallGenerationModelSettings, type ToolCallGenerationStartedEvent, ToolCallParseError, type ToolCallPart, type ToolCallPromptTemplate, type ToolCallResult, type ToolCallsGenerationFinishedEvent, type ToolCallsGenerationFinishedEventResult, type ToolCallsGenerationModel, type ToolCallsGenerationModelSettings, type ToolCallsGenerationStartedEvent, ToolCallsParseError, type ToolCallsPromptTemplate, type ToolContent, type ToolDefinition, ToolExecutionError, type ToolResponsePart, type TranscriptionFinishedEvent, type TranscriptionFinishedEventResult, type TranscriptionModel, type TranscriptionModelSettings, type TranscriptionStartedEvent, TypeValidationError, UncheckedSchema, type UpsertIntoVectorIndexFinishedEvent, type UpsertIntoVectorIndexStartedEvent, type UrlParts, type UserContent, type ValueCluster, type Vector, type VectorIndex, VectorIndexRetriever, type VectorIndexRetrieverSettings, VicunaPromptTemplate as VicunaPrompt, WebSearchTool, type WebSearchToolInput, type WebSearchToolOutput, WhisperCppApiConfiguration, WhisperCppTranscriptionModel, type WhisperCppTranscriptionModelSettings, ZodSchema, ApiFacade as api, Automatic1111Facade as automatic1111, calculateOpenAIChatCostInMillicents, calculateOpenAICompletionCostInMillicents, calculateOpenAIEmbeddingCostInMillicents, calculateOpenAIImageGenerationCostInMillicents, calculateOpenAISpeechCostInMillicents, calculateOpenAITranscriptionCostInMillicents, classify, CohereFacade as cohere, convertDataContentToBase64String, convertDataContentToUint8Array, cosineSimilarity, countOpenAIChatMessageTokens, countOpenAIChatPromptTokens, countTokens, createChatPrompt, createEventSourceStream, createInstructionPrompt, createTextPrompt, delay, ElevenLabsFacade as elevenlabs, embed, embedMany, executeFunction, executeTool, generateImage, generateObject, generateSpeech, generateText, generateToolCall, generateToolCalls, generateTranscription, getAudioFileExtension, getOpenAIChatModelInformation, getOpenAICompletionModelInformation, getRun, HuggingFaceFacade as huggingface, isOpenAIChatModel, isOpenAICompletionModel, isOpenAIEmbeddingModel, isPromptFunction, jsonObjectPrompt, jsonToolCallPrompt, LlamaCppFacade as llamacpp, LmntFacade as lmnt, mapBasicPromptToAutomatic1111Format, mapBasicPromptToStabilityFormat, markAsPromptFunction, MistralFacade as mistral, ModelFusionConfiguration as modelfusion, OllamaFacade as ollama, OpenAIFacade as openai, OpenAICompatibleFacade as openaicompatible, parseJSON, retrieve, retryNever, retryWithExponentialBackoff, runTool, type runToolFinishedEvent, type runToolStartedEvent, runTools, type runToolsFinishedEvent, type runToolsStartedEvent, safeParseJSON, safeValidateTypes, splitAtCharacter, splitAtToken, splitOnSeparator, splitTextChunk, splitTextChunks, StabilityFacade as stability, streamObject, streamSpeech, streamText, textGenerationModelProperties, throttleMaxConcurrency, throttleOff, trimChatPrompt, uncheckedSchema, upsertIntoVectorIndex, validateContentIsString, validateTypes, WhisperCppFacade as whispercpp, withRun, zodSchema };
         | 
| 9462 | 
            +
            export { AbortError, AbstractOpenAIChatModel, type AbstractOpenAIChatSettings, AbstractOpenAICompletionModel, type AbstractOpenAICompletionModelSettings, AbstractOpenAITextEmbeddingModel, type AbstractOpenAITextEmbeddingModelSettings, AlpacaPromptTemplate as AlpacaPrompt, ApiCallError, type ApiConfiguration, type AssistantContent, AsyncQueue, type AudioMimeType, Automatic1111ApiConfiguration, type Automatic1111ErrorData, Automatic1111ImageGenerationModel, type Automatic1111ImageGenerationPrompt, type Automatic1111ImageGenerationResponse, type Automatic1111ImageGenerationSettings, AzureOpenAIApiConfiguration, type AzureOpenAIApiConfigurationOptions, type BaseFunctionEvent, type BaseFunctionFinishedEvent, type BaseFunctionFinishedEventResult, type BaseFunctionStartedEvent, type BaseModelCallFinishedEvent, type BaseModelCallFinishedEventResult, type BaseModelCallStartedEvent, BaseUrlApiConfiguration, BaseUrlApiConfigurationWithDefaults, type BaseUrlPartsApiConfigurationOptions, type BasicTokenizer, CHAT_MODEL_CONTEXT_WINDOW_SIZES, COHERE_TEXT_EMBEDDING_MODELS, COHERE_TEXT_GENERATION_MODELS, type Cache, ChatMLPromptTemplate as ChatMLPrompt, ChatMessage, type ChatPrompt, type Classifier, type ClassifierSettings, type ClassifyFinishedEvent, type ClassifyFinishedEventResult, type ClassifyStartedEvent, CohereApiConfiguration, type CohereDetokenizationResponse, type CohereErrorData, CohereTextEmbeddingModel, type CohereTextEmbeddingModelSettings, type CohereTextEmbeddingModelType, type CohereTextEmbeddingResponse, CohereTextGenerationModel, type CohereTextGenerationModelSettings, type CohereTextGenerationModelType, type CohereTextGenerationResponse, CohereTextGenerationResponseFormat, type CohereTextGenerationResponseFormatType, type CohereTextStreamChunk, type CohereTokenizationResponse, CohereTokenizer, type CohereTokenizerModelType, type CohereTokenizerSettings, type CustomHeaderProvider, type DataContent, DefaultRun, type Delta, ElevenLabsApiConfiguration, ElevenLabsSpeechModel, type ElevenLabsSpeechModelSettings, type EmbeddingFinishedEvent, type EmbeddingFinishedEventResult, type EmbeddingModel, type EmbeddingModelSettings, EmbeddingSimilarityClassifier, type EmbeddingSimilarityClassifierSettings, type EmbeddingStartedEvent, type ExecuteFunctionFinishedEvent, type ExecuteFunctionStartedEvent, type ExecuteToolFinishedEvent, type ExecuteToolMetadata, type ExecuteToolStartedEvent, FireworksAIApiConfiguration, type FlexibleObjectFromTextPromptTemplate, type FullTokenizer, type FunctionCallOptions, type FunctionEvent, FunctionEventSource, type FunctionObserver, type FunctionOptions, type HasContextWindowSize, type HasTokenizer, type HeaderParameters, HeliconeOpenAIApiConfiguration, HuggingFaceApiConfiguration, type HuggingFaceErrorData, HuggingFaceTextEmbeddingModel, type HuggingFaceTextEmbeddingModelSettings, type HuggingFaceTextEmbeddingResponse, HuggingFaceTextGenerationModel, type HuggingFaceTextGenerationModelSettings, type HuggingFaceTextGenerationResponse, type ImageGenerationFinishedEvent, type ImageGenerationFinishedEventResult, type ImageGenerationModel, type ImageGenerationModelSettings, type ImageGenerationStartedEvent, type ImagePart, type InstructionContent, type InstructionPrompt, InvalidPromptError, JSONParseError, type JsonSchemaProducer, Llama2PromptTemplate as Llama2Prompt, LlamaCppApiConfiguration, LlamaCppCompletionModel, type LlamaCppCompletionModelSettings, type LlamaCppCompletionPrompt, LlamaCppCompletionResponseFormat, type LlamaCppCompletionResponseFormatType, type LlamaCppErrorData, LlamaCppTextEmbeddingModel, type LlamaCppTextEmbeddingModelSettings, type LlamaCppTextEmbeddingResponse, type LlamaCppTextGenerationResponse, type LlamaCppTextStreamChunk, type LlamaCppTokenizationResponse, LlamaCppTokenizer, LmntApiConfiguration, LmntSpeechModel, type LmntSpeechModelSettings, type LmntSpeechResponse, type LogFormat, MemoryCache, MemoryVectorIndex, MistralApiConfiguration, type MistralChatMessage, MistralChatModel, type MistralChatModelSettings, type MistralChatPrompt, type MistralChatResponse, MistralChatResponseFormat, type MistralChatResponseFormatType, type MistralChatStreamChunk, type MistralErrorData, MistralInstructPromptTemplate as MistralInstructPrompt, MistralTextEmbeddingModel, type MistralTextEmbeddingModelSettings, type MistralTextEmbeddingResponse, type Model, type ModelCallFinishedEvent, type ModelCallMetadata, type ModelCallStartedEvent, type ModelInformation, type ModelSettings, NeuralChatPromptTemplate as NeuralChatPrompt, NoSuchToolDefinitionError, OPENAI_CHAT_MESSAGE_BASE_TOKEN_COUNT, OPENAI_CHAT_PROMPT_BASE_TOKEN_COUNT, OPENAI_TEXT_EMBEDDING_MODELS, OPENAI_TEXT_GENERATION_MODELS, ObjectFromTextGenerationModel, type ObjectFromTextPromptTemplate, ObjectFromTextStreamingModel, type ObjectGenerationFinishedEvent, type ObjectGenerationFinishedEventResult, type ObjectGenerationModel, type ObjectGenerationModelSettings, type ObjectGenerationStartedEvent, ObjectGeneratorTool, ObjectParseError, type ObjectStream, ObjectStreamFromResponse, ObjectStreamResponse, type ObjectStreamingFinishedEvent, type ObjectStreamingModel, type ObjectStreamingStartedEvent, ObjectValidationError, OllamaApiConfiguration, type OllamaChatMessage, OllamaChatModel, type OllamaChatModelSettings, type OllamaChatPrompt, type OllamaChatResponse, OllamaChatResponseFormat, type OllamaChatResponseFormatType, type OllamaChatStreamChunk, OllamaCompletionModel, type OllamaCompletionModelSettings, type OllamaCompletionPrompt$1 as OllamaCompletionPrompt, type OllamaCompletionResponse, OllamaCompletionResponseFormat, type OllamaCompletionResponseFormatType, type OllamaCompletionStreamChunk, type OllamaErrorData, OllamaTextEmbeddingModel, type OllamaTextEmbeddingModelSettings, type OllamaTextEmbeddingResponse, type OllamaTextGenerationSettings, OpenAIApiConfiguration, type OpenAIChatBaseModelType, type OpenAIChatChunk, OpenAIChatMessage, OpenAIChatModel, type OpenAIChatModelType, type OpenAIChatPrompt, type OpenAIChatResponse, OpenAIChatResponseFormat, type OpenAIChatResponseFormatType, type OpenAIChatSettings, type OpenAICompatibleApiConfiguration, OpenAICompatibleChatModel, type OpenAICompatibleChatSettings, OpenAICompatibleCompletionModel, type OpenAICompatibleCompletionModelSettings, type OpenAICompatibleProviderName, OpenAICompatibleTextEmbeddingModel, type OpenAICompatibleTextEmbeddingModelSettings, OpenAICompletionModel, type OpenAICompletionModelSettings, type OpenAICompletionModelType, type OpenAICompletionResponse, type OpenAIErrorData, type OpenAIImageGenerationBase64JsonResponse, type OpenAIImageGenerationCallSettings, OpenAIImageGenerationModel, OpenAIImageGenerationResponseFormat, type OpenAIImageGenerationResponseFormatType, type OpenAIImageGenerationSettings, type OpenAIImageGenerationUrlResponse, OpenAISpeechModel, type OpenAISpeechModelSettings, type OpenAISpeechModelType, type OpenAISpeechVoice, OpenAITextEmbeddingModel, type OpenAITextEmbeddingModelSettings, type OpenAITextEmbeddingModelType, type OpenAITextEmbeddingResponse, OpenAITextResponseFormat, type OpenAITextResponseFormatType, type OpenAITranscriptionJsonResponse, OpenAITranscriptionModel, type OpenAITranscriptionModelSettings, OpenAITranscriptionResponseFormat, type OpenAITranscriptionResponseFormatType, type OpenAITranscriptionVerboseJsonResponse, type PartialBaseUrlPartsApiConfigurationOptions, PerplexityApiConfiguration, type PromptFunction, type PromptTemplate, PromptTemplateFullTextModel, PromptTemplateImageGenerationModel, PromptTemplateTextGenerationModel, PromptTemplateTextStreamingModel, type Retriever, RetryError, type RetryErrorReason, type RetryFunction, type Run, type Schema, type SpeechGenerationFinishedEvent, type SpeechGenerationFinishedEventResult, type SpeechGenerationModel, type SpeechGenerationModelSettings, type SpeechGenerationStartedEvent, type SpeechStreamingFinishedEvent, type SpeechStreamingStartedEvent, type SplitFunction, StabilityApiConfiguration, type StabilityClipGuidancePreset, type StabilityErrorData, StabilityImageGenerationModel, type StabilityImageGenerationModelType, type StabilityImageGenerationPrompt, type StabilityImageGenerationResponse, type StabilityImageGenerationSampler, type StabilityImageGenerationSettings, type StabilityImageGenerationStylePreset, type StreamingSpeechGenerationModel, SynthiaPromptTemplate as SynthiaPrompt, type TextChunk, type TextGenerationBaseModel, type TextGenerationFinishReason, type TextGenerationFinishedEvent, type TextGenerationFinishedEventResult, type TextGenerationModel, type TextGenerationModelSettings, type TextGenerationPromptTemplate, type TextGenerationPromptTemplateProvider, type TextGenerationResult, type TextGenerationStartedEvent, TextGenerationToolCallModel, TextGenerationToolCallsModel, type TextPart, TextPromptTemplate as TextPrompt, type TextStreamingBaseModel, type TextStreamingFinishedEvent, type TextStreamingModel, type TextStreamingStartedEvent, type ThrottleFunction, TikTokenTokenizer, type TikTokenTokenizerSettings, TogetherAIApiConfiguration, Tool, type ToolCall, ToolCallArgumentsValidationError, ToolCallError, ToolCallGenerationError, type ToolCallGenerationFinishedEvent, type ToolCallGenerationFinishedEventResult, type ToolCallGenerationModel, type ToolCallGenerationModelSettings, type ToolCallGenerationStartedEvent, ToolCallParseError, type ToolCallPart, type ToolCallPromptTemplate, type ToolCallResult, type ToolCallsGenerationFinishedEvent, type ToolCallsGenerationFinishedEventResult, type ToolCallsGenerationModel, type ToolCallsGenerationModelSettings, type ToolCallsGenerationStartedEvent, ToolCallsParseError, type ToolCallsPromptTemplate, type ToolContent, type ToolDefinition, ToolExecutionError, type ToolResponsePart, type TranscriptionFinishedEvent, type TranscriptionFinishedEventResult, type TranscriptionModel, type TranscriptionModelSettings, type TranscriptionStartedEvent, TypeValidationError, UncheckedSchema, type UpsertIntoVectorIndexFinishedEvent, type UpsertIntoVectorIndexStartedEvent, type UrlParts, type UserContent, type ValueCluster, type Vector, type VectorIndex, VectorIndexRetriever, type VectorIndexRetrieverSettings, VicunaPromptTemplate as VicunaPrompt, WebSearchTool, type WebSearchToolInput, type WebSearchToolOutput, WhisperCppApiConfiguration, WhisperCppTranscriptionModel, type WhisperCppTranscriptionModelSettings, ZodSchema, ApiFacade as api, Automatic1111Facade as automatic1111, classify, CohereFacade as cohere, convertDataContentToBase64String, convertDataContentToUint8Array, cosineSimilarity, countOpenAIChatMessageTokens, countOpenAIChatPromptTokens, countTokens, createChatPrompt, createEventSourceStream, createInstructionPrompt, createTextPrompt, delay, ElevenLabsFacade as elevenlabs, embed, embedMany, executeFunction, executeTool, generateImage, generateObject, generateSpeech, generateText, generateToolCall, generateToolCalls, generateTranscription, getAudioFileExtension, getOpenAIChatModelInformation, getOpenAICompletionModelInformation, getRun, HuggingFaceFacade as huggingface, isPromptFunction, jsonObjectPrompt, jsonToolCallPrompt, LlamaCppFacade as llamacpp, LmntFacade as lmnt, mapBasicPromptToAutomatic1111Format, mapBasicPromptToStabilityFormat, markAsPromptFunction, MistralFacade as mistral, ModelFusionConfiguration as modelfusion, OllamaFacade as ollama, openAITextEmbeddingResponseSchema, OpenAIFacade as openai, OpenAICompatibleFacade as openaicompatible, parseJSON, retrieve, retryNever, retryWithExponentialBackoff, runTool, type runToolFinishedEvent, type runToolStartedEvent, runTools, type runToolsFinishedEvent, type runToolsStartedEvent, safeParseJSON, safeValidateTypes, splitAtCharacter, splitAtToken, splitOnSeparator, splitTextChunk, splitTextChunks, StabilityFacade as stability, streamObject, streamSpeech, streamText, textGenerationModelProperties, throttleMaxConcurrency, throttleOff, trimChatPrompt, uncheckedSchema, upsertIntoVectorIndex, validateContentIsString, validateTypes, WhisperCppFacade as whispercpp, withRun, zodSchema };
         |