@budibase/pro 3.23.35 → 3.23.37
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/ai/index.d.ts +1 -0
- package/dist/ai/models/litellm.d.ts +15 -0
- package/dist/ai/prompts/index.d.ts +7 -0
- package/dist/index.js +48 -27
- package/dist/sdk/quotas/quotas.d.ts +1 -1
- package/dist/types/ai.d.ts +7 -0
- package/package.json +2 -2
package/dist/ai/index.d.ts
CHANGED
|
@@ -3,6 +3,7 @@ export * from "./prompts";
|
|
|
3
3
|
export { parseResponseFormat } from "./models/openai";
|
|
4
4
|
export { getLLM, getLLMConfig, getLLMOrThrow, LLMRequest, getChatLLM, } from "./llm";
|
|
5
5
|
export { LLM } from "./models/base";
|
|
6
|
+
export { createLiteLLMOpenAI, getLiteLLMProvider, getLiteLLMProviderOptions, } from "./models/litellm";
|
|
6
7
|
export * from "./prompts";
|
|
7
8
|
export * from "./generators";
|
|
8
9
|
export * from "./structuredOutputs";
|
|
@@ -1,6 +1,21 @@
|
|
|
1
1
|
import { LLMConfigOptions } from "@budibase/types";
|
|
2
2
|
import { OpenAI } from "./openai";
|
|
3
3
|
import { default as OpenAIClient } from "openai";
|
|
4
|
+
type LiteLLMFetch = (input: Parameters<typeof fetch>[0], init?: Parameters<typeof fetch>[1]) => ReturnType<typeof fetch>;
|
|
5
|
+
type LiteLLMOpenAIConfig = {
|
|
6
|
+
apiKey: string;
|
|
7
|
+
baseUrl: string;
|
|
8
|
+
fetch?: LiteLLMFetch;
|
|
9
|
+
};
|
|
4
10
|
export declare class LiteLLMAI extends OpenAI {
|
|
5
11
|
protected getClient(opts: LLMConfigOptions): OpenAIClient;
|
|
6
12
|
}
|
|
13
|
+
export declare const getLiteLLMProvider: (modelId: string) => string;
|
|
14
|
+
export declare const getLiteLLMProviderOptions: (modelId: string) => {
|
|
15
|
+
parallelToolCalls: true;
|
|
16
|
+
} | {
|
|
17
|
+
reasoningEffort: "low";
|
|
18
|
+
parallelToolCalls: true;
|
|
19
|
+
};
|
|
20
|
+
export declare const createLiteLLMOpenAI: (config: LiteLLMOpenAIConfig) => import("@ai-sdk/openai").OpenAIProvider;
|
|
21
|
+
export {};
|
|
@@ -1,5 +1,10 @@
|
|
|
1
1
|
import { EnrichedBinding, Snippet, ContextUser, SummariseLength } from "@budibase/types";
|
|
2
2
|
import { LLMRequest } from "../llm";
|
|
3
|
+
import { AgentPromptOptions } from "../../types";
|
|
4
|
+
export interface AutomationAgentToolGuideline {
|
|
5
|
+
toolName: string;
|
|
6
|
+
guidelines: string;
|
|
7
|
+
}
|
|
3
8
|
export declare function summarizeText(text: string, length?: SummariseLength): LLMRequest;
|
|
4
9
|
export declare function extractFileData(schema: Record<string, any>, fileIdOrDataUrl: string): LLMRequest;
|
|
5
10
|
export declare function classifyText(text: string, categories: string[]): LLMRequest;
|
|
@@ -14,5 +19,7 @@ export declare function generateJs(bindings: EnrichedBinding[], snippets: Snippe
|
|
|
14
19
|
export declare function generateTables(): LLMRequest;
|
|
15
20
|
export declare function generateAIColumns(): LLMRequest;
|
|
16
21
|
export declare function generateData(): LLMRequest;
|
|
22
|
+
export declare function composeAutomationAgentSystemPrompt(options: AgentPromptOptions): string;
|
|
23
|
+
export declare function composeAutomationAgentToolGuidelines(guidelines: AutomationAgentToolGuideline[]): string;
|
|
17
24
|
export declare function agentSystemPrompt(user: ContextUser): string;
|
|
18
25
|
export declare function agentHistoryTitleSystemPrompt(): string;
|