@promptbook/browser 0.66.0-5 → 0.66.0-7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (20) hide show
  1. package/esm/index.es.js +1 -1
  2. package/esm/typings/src/_packages/cli.index.d.ts +6 -2
  3. package/esm/typings/src/cli/main.d.ts +2 -2
  4. package/esm/typings/src/execution/LlmExecutionTools.d.ts +1 -0
  5. package/esm/typings/src/knowledge/prepare-knowledge/_common/prepareKnowledgePieces.test.d.ts +1 -1
  6. package/esm/typings/src/knowledge/prepare-knowledge/markdown/prepareKnowledgeFromMarkdown.test.d.ts +1 -1
  7. package/esm/typings/src/knowledge/prepare-knowledge/pdf/prepareKnowledgeFromPdf.test.d.ts +1 -1
  8. package/esm/typings/src/llm-providers/_common/getLlmToolsForTestingAndScriptsAndPlayground.d.ts +1 -0
  9. package/esm/typings/src/llm-providers/anthropic-claude/AnthropicClaudeExecutionTools.d.ts +10 -5
  10. package/esm/typings/src/llm-providers/anthropic-claude/register-configuration.d.ts +1 -0
  11. package/esm/typings/src/llm-providers/azure-openai/AzureOpenAiExecutionTools.d.ts +10 -5
  12. package/esm/typings/src/llm-providers/mocked/MockedEchoLlmExecutionTools.d.ts +8 -4
  13. package/esm/typings/src/llm-providers/mocked/MockedFackedLlmExecutionTools.d.ts +8 -4
  14. package/esm/typings/src/llm-providers/multiple/MultipleLlmExecutionTools.d.ts +9 -5
  15. package/esm/typings/src/llm-providers/openai/OpenAiExecutionTools.d.ts +10 -5
  16. package/esm/typings/src/llm-providers/openai/register-configuration.d.ts +1 -0
  17. package/esm/typings/src/llm-providers/remote/RemoteLlmExecutionTools.d.ts +8 -4
  18. package/esm/typings/src/personas/preparePersona.test.d.ts +1 -1
  19. package/package.json +2 -2
  20. package/umd/index.umd.js +1 -1
package/esm/index.es.js CHANGED
@@ -5,7 +5,7 @@ import { isRunningInBrowser } from 'openai/core';
5
5
  /**
6
6
  * The version of the Promptbook library
7
7
  */
8
- var PROMPTBOOK_VERSION = '0.66.0-4';
8
+ var PROMPTBOOK_VERSION = '0.66.0-6';
9
9
  // TODO: !!!! List here all the versions and annotate + put into script
10
10
 
11
11
  /*! *****************************************************************************
@@ -1,4 +1,8 @@
1
1
  import { PROMPTBOOK_VERSION } from '../version';
2
- import { __CLI } from '../cli/main';
2
+ import { _CLI } from '../cli/main';
3
+ import { _AnthropicClaudeMetadataRegistration } from '../llm-providers/anthropic-claude/register-configuration';
4
+ import { _OpenAiMetadataRegistration } from '../llm-providers/openai/register-configuration';
3
5
  export { PROMPTBOOK_VERSION };
4
- export { __CLI };
6
+ export { _CLI };
7
+ export { _AnthropicClaudeMetadataRegistration };
8
+ export { _OpenAiMetadataRegistration };
@@ -4,8 +4,8 @@ import { promptbookCli } from './promptbookCli';
4
4
  *
5
5
  * @public exported from `@promptbook/cli`
6
6
  */
7
- export declare const __CLI: {
8
- __initialize: typeof promptbookCli;
7
+ export declare const _CLI: {
8
+ _initialize: typeof promptbookCli;
9
9
  };
10
10
  /**
11
11
  * Note: [🟡] This code should never be published outside of `@promptbook/cli`
@@ -33,6 +33,7 @@ export type LlmExecutionTools = {
33
33
  * @returns nothing if configuration is correct
34
34
  * @throws {Error} if configuration is incorrect
35
35
  */
36
+ checkConfiguration(): Promisable<void>;
36
37
  /**
37
38
  * List all available models that can be used
38
39
  */
@@ -1 +1 @@
1
- import '../../../_packages/core.index';
1
+ export {};
@@ -1,4 +1,4 @@
1
- import '../../../_packages/core.index';
1
+ export {};
2
2
  /**
3
3
  * TODO: [📓] Maybe test all file in samples (not just 10-simple.md)
4
4
  */
@@ -1 +1 @@
1
- import '../../../_packages/core.index';
1
+ export {};
@@ -1,3 +1,4 @@
1
+ import '../../_packages/core.index';
1
2
  import type { CreateLlmToolsFromConfigurationOptions } from './createLlmToolsFromConfiguration';
2
3
  import type { LlmExecutionToolsWithTotalUsage } from './utils/count-total-usage/LlmExecutionToolsWithTotalUsage';
3
4
  type GetLlmToolsForTestingAndScriptsAndPlaygroundOptions = CreateLlmToolsFromConfigurationOptions & {
@@ -17,7 +17,7 @@ export declare class AnthropicClaudeExecutionTools implements LlmExecutionTools
17
17
  /**
18
18
  * Anthropic Claude API client.
19
19
  */
20
- private readonly client;
20
+ private client;
21
21
  /**
22
22
  * Creates Anthropic Claude Execution Tools.
23
23
  *
@@ -26,6 +26,15 @@ export declare class AnthropicClaudeExecutionTools implements LlmExecutionTools
26
26
  constructor(options?: AnthropicClaudeExecutionToolsDirectOptions);
27
27
  get title(): string_title & string_markdown_text;
28
28
  get description(): string_markdown;
29
+ private getClient;
30
+ /**
31
+ * Check the `options` passed to `constructor`
32
+ */
33
+ checkConfiguration(): Promise<void>;
34
+ /**
35
+ * List all available Anthropic Claude models that can be used
36
+ */
37
+ listModels(): Array<AvailableModel>;
29
38
  /**
30
39
  * Calls Anthropic Claude API to use a chat model.
31
40
  */
@@ -38,10 +47,6 @@ export declare class AnthropicClaudeExecutionTools implements LlmExecutionTools
38
47
  * Default model for chat variant.
39
48
  */
40
49
  private getDefaultChatModel;
41
- /**
42
- * List all available Anthropic Claude models that can be used
43
- */
44
- listModels(): Array<AvailableModel>;
45
50
  }
46
51
  /**
47
52
  * TODO: [🍆] JSON mode
@@ -4,5 +4,6 @@
4
4
  * Note: [🏐] Configurations registrations are done in @@@ BUT constructor @@@
5
5
  *
6
6
  * @public exported from `@promptbook/core`
7
+ * @public exported from `@promptbook/cli`
7
8
  */
8
9
  export declare const _AnthropicClaudeMetadataRegistration: void;
@@ -17,7 +17,7 @@ export declare class AzureOpenAiExecutionTools implements LlmExecutionTools {
17
17
  /**
18
18
  * OpenAI Azure API client.
19
19
  */
20
- private readonly client;
20
+ private client;
21
21
  /**
22
22
  * Creates OpenAI Execution Tools.
23
23
  *
@@ -26,6 +26,15 @@ export declare class AzureOpenAiExecutionTools implements LlmExecutionTools {
26
26
  constructor(options: AzureOpenAiExecutionToolsOptions);
27
27
  get title(): string_title & string_markdown_text;
28
28
  get description(): string_markdown;
29
+ private getClient;
30
+ /**
31
+ * Check the `options` passed to `constructor`
32
+ */
33
+ checkConfiguration(): Promise<void>;
34
+ /**
35
+ * List all available Azure OpenAI models that can be used
36
+ */
37
+ listModels(): Promise<Array<AvailableModel>>;
29
38
  /**
30
39
  * Calls OpenAI API to use a chat model.
31
40
  */
@@ -38,10 +47,6 @@ export declare class AzureOpenAiExecutionTools implements LlmExecutionTools {
38
47
  * Changes Azure error (which is not propper Error but object) to propper Error
39
48
  */
40
49
  private transformAzureError;
41
- /**
42
- * List all available Azure OpenAI models that can be used
43
- */
44
- listModels(): Promise<Array<AvailableModel>>;
45
50
  }
46
51
  /**
47
52
  * TODO: Maybe Create some common util for callChatModel and callCompletionModel
@@ -17,6 +17,14 @@ export declare class MockedEchoLlmExecutionTools implements LlmExecutionTools {
17
17
  constructor(options?: CommonExecutionToolsOptions);
18
18
  get title(): string_title & string_markdown_text;
19
19
  get description(): string_markdown;
20
+ /**
21
+ * Does nothing, just to implement the interface
22
+ */
23
+ checkConfiguration(): void;
24
+ /**
25
+ * List all available mocked-models that can be used
26
+ */
27
+ listModels(): Array<AvailableModel>;
20
28
  /**
21
29
  * Mocks chat model
22
30
  */
@@ -25,10 +33,6 @@ export declare class MockedEchoLlmExecutionTools implements LlmExecutionTools {
25
33
  * Mocks completion model
26
34
  */
27
35
  callCompletionModel(prompt: Pick<Prompt, 'content' | 'parameters' | 'modelRequirements'>): Promise<CompletionPromptResult>;
28
- /**
29
- * List all available mocked-models that can be used
30
- */
31
- listModels(): Array<AvailableModel>;
32
36
  }
33
37
  /**
34
38
  * TODO: [🧠][🈁] Maybe use `isDeterministic` from options
@@ -18,6 +18,14 @@ export declare class MockedFackedLlmExecutionTools implements LlmExecutionTools
18
18
  constructor(options?: CommonExecutionToolsOptions);
19
19
  get title(): string_title & string_markdown_text;
20
20
  get description(): string_markdown;
21
+ /**
22
+ * Does nothing, just to implement the interface
23
+ */
24
+ checkConfiguration(): void;
25
+ /**
26
+ * List all available fake-models that can be used
27
+ */
28
+ listModels(): Array<AvailableModel>;
21
29
  /**
22
30
  * Fakes chat model
23
31
  */
@@ -30,10 +38,6 @@ export declare class MockedFackedLlmExecutionTools implements LlmExecutionTools
30
38
  * Fakes embedding model
31
39
  */
32
40
  callEmbeddingModel(prompt: Pick<Prompt, 'content' | 'parameters' | 'modelRequirements' | 'expectations' | 'postprocessing'>): Promise<EmbeddingPromptResult>;
33
- /**
34
- * List all available fake-models that can be used
35
- */
36
- listModels(): Array<AvailableModel>;
37
41
  }
38
42
  /**
39
43
  * TODO: [🧠][🈁] Maybe use `isDeterministic` from options
@@ -28,6 +28,15 @@ export declare class MultipleLlmExecutionTools implements LlmExecutionTools {
28
28
  constructor(...llmExecutionTools: Array<LlmExecutionTools>);
29
29
  get title(): string_title & string_markdown_text;
30
30
  get description(): string_markdown;
31
+ /**
32
+ * Check the configuration of all execution tools
33
+ */
34
+ checkConfiguration(): Promise<void>;
35
+ /**
36
+ * List all available models that can be used
37
+ * This lists is a combination of all available models from all execution tools
38
+ */
39
+ listModels(): Promise<Array<AvailableModel>>;
31
40
  /**
32
41
  * Calls the best available chat model
33
42
  */
@@ -46,11 +55,6 @@ export declare class MultipleLlmExecutionTools implements LlmExecutionTools {
46
55
  * Note: This should be private or protected but is public to be usable with duck typing
47
56
  */
48
57
  callCommonModel(prompt: Prompt): Promise<PromptResult>;
49
- /**
50
- * List all available models that can be used
51
- * This lists is a combination of all available models from all execution tools
52
- */
53
- listModels(): Promise<Array<AvailableModel>>;
54
58
  }
55
59
  /**
56
60
  * TODO: [🧠][🎛] Aggregating multiple models - have result not only from one first aviable model BUT all of them
@@ -18,7 +18,7 @@ export declare class OpenAiExecutionTools implements LlmExecutionTools {
18
18
  /**
19
19
  * OpenAI API client.
20
20
  */
21
- private readonly client;
21
+ private client;
22
22
  /**
23
23
  * Creates OpenAI Execution Tools.
24
24
  *
@@ -27,6 +27,15 @@ export declare class OpenAiExecutionTools implements LlmExecutionTools {
27
27
  constructor(options?: OpenAiExecutionToolsOptions);
28
28
  get title(): string_title & string_markdown_text;
29
29
  get description(): string_markdown;
30
+ private getClient;
31
+ /**
32
+ * Check the `options` passed to `constructor`
33
+ */
34
+ checkConfiguration(): Promise<void>;
35
+ /**
36
+ * List all available OpenAI models that can be used
37
+ */
38
+ listModels(): Array<AvailableModel>;
30
39
  /**
31
40
  * Calls OpenAI API to use a chat model.
32
41
  */
@@ -55,10 +64,6 @@ export declare class OpenAiExecutionTools implements LlmExecutionTools {
55
64
  * Default model for completion variant.
56
65
  */
57
66
  private getDefaultEmbeddingModel;
58
- /**
59
- * List all available OpenAI models that can be used
60
- */
61
- listModels(): Array<AvailableModel>;
62
67
  }
63
68
  /**
64
69
  * TODO: [🧠][🧙‍♂️] Maybe there can be some wizzard for thoose who want to use just OpenAI
@@ -4,5 +4,6 @@
4
4
  * Note: [🏐] Configurations registrations are done in @@@ BUT constructor @@@
5
5
  *
6
6
  * @public exported from `@promptbook/core`
7
+ * @public exported from `@promptbook/cli`
7
8
  */
8
9
  export declare const _OpenAiMetadataRegistration: void;
@@ -24,6 +24,14 @@ export declare class RemoteLlmExecutionTools implements LlmExecutionTools {
24
24
  constructor(options: RemoteLlmExecutionToolsOptions);
25
25
  get title(): string_title & string_markdown_text;
26
26
  get description(): string_markdown;
27
+ /**
28
+ * Check the configuration of all execution tools
29
+ */
30
+ checkConfiguration(): Promise<void>;
31
+ /**
32
+ * List all available models that can be used
33
+ */
34
+ listModels(): Promise<Array<AvailableModel>>;
27
35
  /**
28
36
  * Creates a connection to the remote proxy server.
29
37
  */
@@ -44,10 +52,6 @@ export declare class RemoteLlmExecutionTools implements LlmExecutionTools {
44
52
  * Calls remote proxy server to use both completion or chat model
45
53
  */
46
54
  private callCommonModel;
47
- /**
48
- * List all available models that can be used
49
- */
50
- listModels(): Promise<Array<AvailableModel>>;
51
55
  }
52
56
  /**
53
57
  * TODO: [🍓] Allow to list compatible models with each variant
@@ -1 +1 @@
1
- import '../../src/_packages/core.index';
1
+ export {};
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@promptbook/browser",
3
- "version": "0.66.0-5",
3
+ "version": "0.66.0-7",
4
4
  "description": "Supercharge your use of large language models",
5
5
  "private": false,
6
6
  "sideEffects": false,
@@ -47,7 +47,7 @@
47
47
  "module": "./esm/index.es.js",
48
48
  "typings": "./esm/typings/src/_packages/browser.index.d.ts",
49
49
  "peerDependencies": {
50
- "@promptbook/core": "0.66.0-5"
50
+ "@promptbook/core": "0.66.0-7"
51
51
  },
52
52
  "dependencies": {
53
53
  "spacetrim": "0.11.39"
package/umd/index.umd.js CHANGED
@@ -8,7 +8,7 @@
8
8
  /**
9
9
  * The version of the Promptbook library
10
10
  */
11
- var PROMPTBOOK_VERSION = '0.66.0-4';
11
+ var PROMPTBOOK_VERSION = '0.66.0-6';
12
12
  // TODO: !!!! List here all the versions and annotate + put into script
13
13
 
14
14
  /*! *****************************************************************************