@promptbook/openai 0.63.1 → 0.63.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/esm/index.es.js CHANGED
@@ -6,7 +6,7 @@ import spaceTrim$1, { spaceTrim } from 'spacetrim';
6
6
  /**
7
7
  * The version of the Promptbook library
8
8
  */
9
- var PROMPTBOOK_VERSION = '0.63.0';
9
+ var PROMPTBOOK_VERSION = '0.63.2';
10
10
  // TODO: !!!! List here all the versions and annotate + put into script
11
11
 
12
12
  /*! *****************************************************************************
@@ -36,6 +36,7 @@ import type { LlmExecutionToolsWithTotalUsage } from '../llm-providers/_common/u
36
36
  import type { AnthropicClaudeExecutionToolsOptions } from '../llm-providers/anthropic-claude/AnthropicClaudeExecutionToolsOptions';
37
37
  import type { AzureOpenAiExecutionToolsOptions } from '../llm-providers/azure-openai/AzureOpenAiExecutionToolsOptions';
38
38
  import type { LangtailExecutionToolsOptions } from '../llm-providers/langtail/LangtailExecutionToolsOptions';
39
+ import type { MultipleLlmExecutionTools } from '../llm-providers/multiple/MultipleLlmExecutionTools';
39
40
  import type { OpenAiExecutionToolsOptions } from '../llm-providers/openai/OpenAiExecutionToolsOptions';
40
41
  import type { Promptbook_Server_Error } from '../llm-providers/remote/interfaces/Promptbook_Server_Error';
41
42
  import type { Promptbook_Server_Progress } from '../llm-providers/remote/interfaces/Promptbook_Server_Progress';
@@ -242,6 +243,7 @@ export type { LlmExecutionToolsWithTotalUsage };
242
243
  export type { AnthropicClaudeExecutionToolsOptions };
243
244
  export type { AzureOpenAiExecutionToolsOptions };
244
245
  export type { LangtailExecutionToolsOptions };
246
+ export type { MultipleLlmExecutionTools };
245
247
  export type { OpenAiExecutionToolsOptions };
246
248
  export type { Promptbook_Server_Error };
247
249
  export type { Promptbook_Server_Progress };
@@ -3,22 +3,25 @@ import type { LlmExecutionTools } from '../../execution/LlmExecutionTools';
3
3
  import type { ChatPromptResult } from '../../execution/PromptResult';
4
4
  import type { CompletionPromptResult } from '../../execution/PromptResult';
5
5
  import type { EmbeddingPromptResult } from '../../execution/PromptResult';
6
+ import type { PromptResult } from '../../execution/PromptResult';
6
7
  import type { ChatPrompt } from '../../types/Prompt';
7
8
  import type { CompletionPrompt } from '../../types/Prompt';
8
9
  import type { EmbeddingPrompt } from '../../types/Prompt';
10
+ import type { Prompt } from '../../types/Prompt';
9
11
  import type { string_markdown } from '../../types/typeAliases';
10
12
  import type { string_markdown_text } from '../../types/typeAliases';
11
13
  import type { string_title } from '../../types/typeAliases';
12
14
  /**
13
15
  * Multiple LLM Execution Tools is a proxy server that uses multiple execution tools internally and exposes the executor interface externally.
14
16
  *
15
- * @private internal utility of `joinLlmExecutionTools`
17
+ * Note: Internal utility of `joinLlmExecutionTools` but exposed type
18
+ * @public exported from `@promptbook/types`
16
19
  */
17
20
  export declare class MultipleLlmExecutionTools implements LlmExecutionTools {
18
21
  /**
19
22
  * Array of execution tools in order of priority
20
23
  */
21
- private readonly llmExecutionTools;
24
+ readonly llmExecutionTools: Array<LlmExecutionTools>;
22
25
  /**
23
26
  * Gets array of execution tools in order of priority
24
27
  */
@@ -40,7 +43,7 @@ export declare class MultipleLlmExecutionTools implements LlmExecutionTools {
40
43
  /**
41
44
  * Calls the best available model
42
45
  */
43
- private callModelCommon;
46
+ protected callCommonModel(prompt: Prompt): Promise<PromptResult>;
44
47
  /**
45
48
  * List all available models that can be used
46
49
  * This lists is a combination of all available models from all execution tools
@@ -43,7 +43,7 @@ export declare class RemoteLlmExecutionTools implements LlmExecutionTools {
43
43
  /**
44
44
  * Calls remote proxy server to use both completion or chat model
45
45
  */
46
- private callModelCommon;
46
+ private callCommonModel;
47
47
  /**
48
48
  * List all available models that can be used
49
49
  */
@@ -53,4 +53,4 @@ export declare class RemoteLlmExecutionTools implements LlmExecutionTools {
53
53
  * TODO: [🍓] Allow to list compatible models with each variant
54
54
  * TODO: [🗯] RemoteLlmExecutionTools should extend Destroyable and implement IDestroyable
55
55
  * TODO: [🍜] Add anonymous option
56
- */
56
+ */
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@promptbook/openai",
3
- "version": "0.63.1",
3
+ "version": "0.63.3",
4
4
  "description": "Supercharge your use of large language models",
5
5
  "private": false,
6
6
  "sideEffects": false,
@@ -47,7 +47,7 @@
47
47
  "module": "./esm/index.es.js",
48
48
  "typings": "./esm/typings/src/_packages/openai.index.d.ts",
49
49
  "peerDependencies": {
50
- "@promptbook/core": "0.63.1"
50
+ "@promptbook/core": "0.63.3"
51
51
  },
52
52
  "dependencies": {
53
53
  "colors": "1.4.0",
package/umd/index.umd.js CHANGED
@@ -14,7 +14,7 @@
14
14
  /**
15
15
  * The version of the Promptbook library
16
16
  */
17
- var PROMPTBOOK_VERSION = '0.63.0';
17
+ var PROMPTBOOK_VERSION = '0.63.2';
18
18
  // TODO: !!!! List here all the versions and annotate + put into script
19
19
 
20
20
  /*! *****************************************************************************