@promptbook/components 0.104.0-3 → 0.104.0-5
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/esm/index.es.js +23 -2
- package/esm/index.es.js.map +1 -1
- package/esm/typings/src/collection/agent-collection/constructors/agent-collection-in-supabase/AgentsDatabaseSchema.d.ts +18 -15
- package/esm/typings/src/llm-providers/_multiple/MultipleLlmExecutionTools.d.ts +6 -2
- package/esm/typings/src/llm-providers/remote/RemoteLlmExecutionTools.d.ts +1 -0
- package/esm/typings/src/version.d.ts +1 -1
- package/package.json +1 -1
- package/umd/index.umd.js +23 -2
- package/umd/index.umd.js.map +1 -1
|
@@ -3,11 +3,12 @@
|
|
|
3
3
|
* Source of truth: `/apps/agents-server/src/database/schema.sql` *(do not edit table structure here manually)*
|
|
4
4
|
*
|
|
5
5
|
* [💽] Prompt:
|
|
6
|
-
* Re-generate this sub-schema
|
|
6
|
+
* Re-generate this sub-schema from `/apps/agents-server/src/database/schema.ts` *(which was generated from `/apps/agents-server/src/database/migrations/*.sql`)*
|
|
7
|
+
* `AgentsDatabaseSchema` is strict subset of `AgentsServerDatabase`
|
|
7
8
|
* Generate Supabase TypeScript schema which is a subset of `AgentsServerDatabase`
|
|
8
9
|
* containing only tables `Agent` and `AgentHistory`
|
|
9
10
|
*
|
|
10
|
-
* NOTE: This file intentionally omits all other tables (
|
|
11
|
+
* NOTE: This file intentionally omits all other tables (`Metadata`, `ChatHistory`, `ChatFeedback`, `User`, `LlmCache`, etc.)
|
|
11
12
|
* and any extra schemas (e.g. `graphql_public`) to remain a strict subset.
|
|
12
13
|
*/
|
|
13
14
|
export type Json = string | number | boolean | null | {
|
|
@@ -31,6 +32,7 @@ export type AgentsDatabaseSchema = {
|
|
|
31
32
|
preparedModelRequirements: Json | null;
|
|
32
33
|
preparedExternals: Json | null;
|
|
33
34
|
deletedAt: string | null;
|
|
35
|
+
visibility: 'PUBLIC' | 'PRIVATE';
|
|
34
36
|
};
|
|
35
37
|
Insert: {
|
|
36
38
|
id?: number;
|
|
@@ -46,6 +48,7 @@ export type AgentsDatabaseSchema = {
|
|
|
46
48
|
preparedModelRequirements?: Json | null;
|
|
47
49
|
preparedExternals?: Json | null;
|
|
48
50
|
deletedAt?: string | null;
|
|
51
|
+
visibility?: 'PUBLIC' | 'PRIVATE';
|
|
49
52
|
};
|
|
50
53
|
Update: {
|
|
51
54
|
id?: number;
|
|
@@ -61,6 +64,7 @@ export type AgentsDatabaseSchema = {
|
|
|
61
64
|
preparedModelRequirements?: Json | null;
|
|
62
65
|
preparedExternals?: Json | null;
|
|
63
66
|
deletedAt?: string | null;
|
|
67
|
+
visibility?: 'PUBLIC' | 'PRIVATE';
|
|
64
68
|
};
|
|
65
69
|
Relationships: [];
|
|
66
70
|
};
|
|
@@ -92,21 +96,20 @@ export type AgentsDatabaseSchema = {
|
|
|
92
96
|
agentSource?: string;
|
|
93
97
|
promptbookEngineVersion?: string;
|
|
94
98
|
};
|
|
95
|
-
Relationships: [
|
|
99
|
+
Relationships: [
|
|
100
|
+
{
|
|
101
|
+
foreignKeyName: 'AgentHistory_agentName_fkey';
|
|
102
|
+
columns: ['agentName'];
|
|
103
|
+
referencedRelation: 'Agent';
|
|
104
|
+
referencedColumns: ['agentName'];
|
|
105
|
+
}
|
|
106
|
+
];
|
|
96
107
|
};
|
|
97
108
|
};
|
|
98
|
-
Views:
|
|
99
|
-
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
[_ in never]: never;
|
|
103
|
-
};
|
|
104
|
-
Enums: {
|
|
105
|
-
[_ in never]: never;
|
|
106
|
-
};
|
|
107
|
-
CompositeTypes: {
|
|
108
|
-
[_ in never]: never;
|
|
109
|
-
};
|
|
109
|
+
Views: Record<string, never>;
|
|
110
|
+
Functions: Record<string, never>;
|
|
111
|
+
Enums: Record<string, never>;
|
|
112
|
+
CompositeTypes: Record<string, never>;
|
|
110
113
|
};
|
|
111
114
|
};
|
|
112
115
|
type PublicSchema = AgentsDatabaseSchema[Extract<keyof AgentsDatabaseSchema, 'public'>];
|
|
@@ -1,8 +1,8 @@
|
|
|
1
1
|
import type { ChatParticipant } from '../../book-components/Chat/types/ChatParticipant';
|
|
2
2
|
import type { AvailableModel } from '../../execution/AvailableModel';
|
|
3
3
|
import type { LlmExecutionTools } from '../../execution/LlmExecutionTools';
|
|
4
|
-
import type { ChatPromptResult, CompletionPromptResult, EmbeddingPromptResult, PromptResult } from '../../execution/PromptResult';
|
|
5
|
-
import type { ChatPrompt, CompletionPrompt, EmbeddingPrompt, Prompt } from '../../types/Prompt';
|
|
4
|
+
import type { ChatPromptResult, CompletionPromptResult, EmbeddingPromptResult, ImagePromptResult, PromptResult } from '../../execution/PromptResult';
|
|
5
|
+
import type { ChatPrompt, CompletionPrompt, EmbeddingPrompt, ImagePrompt, Prompt } from '../../types/Prompt';
|
|
6
6
|
import type { string_markdown, string_markdown_text, string_title } from '../../types/typeAliases';
|
|
7
7
|
/**
|
|
8
8
|
* Multiple LLM Execution Tools is a proxy server that uses multiple execution tools internally and exposes the executor interface externally.
|
|
@@ -43,6 +43,10 @@ export declare class MultipleLlmExecutionTools implements LlmExecutionTools {
|
|
|
43
43
|
* Calls the best available embedding model
|
|
44
44
|
*/
|
|
45
45
|
callEmbeddingModel(prompt: EmbeddingPrompt): Promise<EmbeddingPromptResult>;
|
|
46
|
+
/**
|
|
47
|
+
* Calls the best available embedding model
|
|
48
|
+
*/
|
|
49
|
+
callImageGenerationModel(prompt: ImagePrompt): Promise<ImagePromptResult>;
|
|
46
50
|
/**
|
|
47
51
|
* Calls the best available model
|
|
48
52
|
*
|
|
@@ -46,6 +46,7 @@ export declare class RemoteLlmExecutionTools<TCustomOptions = undefined> impleme
|
|
|
46
46
|
private callCommonModel;
|
|
47
47
|
}
|
|
48
48
|
/**
|
|
49
|
+
* TODO: !!!! Deprecate pipeline server and all of its components
|
|
49
50
|
* TODO: Maybe use `$exportJson`
|
|
50
51
|
* TODO: [🧠][🛍] Maybe not `isAnonymous: boolean` BUT `mode: 'ANONYMOUS'|'COLLECTION'`
|
|
51
52
|
* TODO: [🍓] Allow to list compatible models with each variant
|
|
@@ -15,7 +15,7 @@ export declare const BOOK_LANGUAGE_VERSION: string_semantic_version;
|
|
|
15
15
|
export declare const PROMPTBOOK_ENGINE_VERSION: string_promptbook_version;
|
|
16
16
|
/**
|
|
17
17
|
* Represents the version string of the Promptbook engine.
|
|
18
|
-
* It follows semantic versioning (e.g., `0.104.0-
|
|
18
|
+
* It follows semantic versioning (e.g., `0.104.0-4`).
|
|
19
19
|
*
|
|
20
20
|
* @generated
|
|
21
21
|
*/
|
package/package.json
CHANGED
package/umd/index.umd.js
CHANGED
|
@@ -30,7 +30,7 @@
|
|
|
30
30
|
* @generated
|
|
31
31
|
* @see https://github.com/webgptorg/promptbook
|
|
32
32
|
*/
|
|
33
|
-
const PROMPTBOOK_ENGINE_VERSION = '0.104.0-
|
|
33
|
+
const PROMPTBOOK_ENGINE_VERSION = '0.104.0-5';
|
|
34
34
|
/**
|
|
35
35
|
* TODO: string_promptbook_version should be constrained to the all versions of Promptbook engine
|
|
36
36
|
* Note: [💞] Ignore a discrepancy between file name and entity name
|
|
@@ -11232,6 +11232,12 @@
|
|
|
11232
11232
|
callEmbeddingModel(prompt) {
|
|
11233
11233
|
return this.callCommonModel(prompt);
|
|
11234
11234
|
}
|
|
11235
|
+
/**
|
|
11236
|
+
* Calls the best available embedding model
|
|
11237
|
+
*/
|
|
11238
|
+
callImageGenerationModel(prompt) {
|
|
11239
|
+
return this.callCommonModel(prompt);
|
|
11240
|
+
}
|
|
11235
11241
|
// <- Note: [🤖]
|
|
11236
11242
|
/**
|
|
11237
11243
|
* Calls the best available model
|
|
@@ -11258,6 +11264,11 @@
|
|
|
11258
11264
|
continue llm;
|
|
11259
11265
|
}
|
|
11260
11266
|
return await llmExecutionTools.callEmbeddingModel(prompt);
|
|
11267
|
+
case 'IMAGE_GENERATION':
|
|
11268
|
+
if (llmExecutionTools.callImageGenerationModel === undefined) {
|
|
11269
|
+
continue llm;
|
|
11270
|
+
}
|
|
11271
|
+
return await llmExecutionTools.callImageGenerationModel(prompt);
|
|
11261
11272
|
// <- case [🤖]:
|
|
11262
11273
|
default:
|
|
11263
11274
|
throw new UnexpectedError(`Unknown model variant "${prompt.modelRequirements.modelVariant}" in ${llmExecutionTools.title}`);
|
|
@@ -12432,6 +12443,15 @@
|
|
|
12432
12443
|
return promptResult;
|
|
12433
12444
|
};
|
|
12434
12445
|
}
|
|
12446
|
+
if (llmTools.callImageGenerationModel !== undefined) {
|
|
12447
|
+
proxyTools.callImageGenerationModel = async (prompt) => {
|
|
12448
|
+
// console.info('[🚕] callImageGenerationModel through countTotalUsage');
|
|
12449
|
+
const promptResult = await llmTools.callImageGenerationModel(prompt);
|
|
12450
|
+
totalUsage = addUsage(totalUsage, promptResult.usage);
|
|
12451
|
+
spending.next(promptResult.usage);
|
|
12452
|
+
return promptResult;
|
|
12453
|
+
};
|
|
12454
|
+
}
|
|
12435
12455
|
// <- Note: [🤖]
|
|
12436
12456
|
return proxyTools;
|
|
12437
12457
|
}
|
|
@@ -13975,8 +13995,9 @@
|
|
|
13975
13995
|
$ongoingTaskResult.$resultString = $ongoingTaskResult.$completionResult.content;
|
|
13976
13996
|
break variant;
|
|
13977
13997
|
case 'EMBEDDING':
|
|
13998
|
+
case 'IMAGE_GENERATION':
|
|
13978
13999
|
throw new PipelineExecutionError(spaceTrim$1.spaceTrim((block) => `
|
|
13979
|
-
|
|
14000
|
+
${modelRequirements.modelVariant} model can not be used in pipeline
|
|
13980
14001
|
|
|
13981
14002
|
This should be catched during parsing
|
|
13982
14003
|
|