@promptbook/markdown-utils 0.92.0-32 → 0.92.0-34

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -102,12 +102,12 @@ import { createLlmToolsFromConfiguration } from '../llm-providers/_common/regist
102
102
  import { cacheLlmTools } from '../llm-providers/_common/utils/cache/cacheLlmTools';
103
103
  import { countUsage } from '../llm-providers/_common/utils/count-total-usage/countUsage';
104
104
  import { limitTotalUsage } from '../llm-providers/_common/utils/count-total-usage/limitTotalUsage';
105
+ import { joinLlmExecutionTools } from '../llm-providers/_multiple/joinLlmExecutionTools';
106
+ import { MultipleLlmExecutionTools } from '../llm-providers/_multiple/MultipleLlmExecutionTools';
105
107
  import { _AnthropicClaudeMetadataRegistration } from '../llm-providers/anthropic-claude/register-configuration';
106
108
  import { _AzureOpenAiMetadataRegistration } from '../llm-providers/azure-openai/register-configuration';
107
109
  import { _DeepseekMetadataRegistration } from '../llm-providers/deepseek/register-configuration';
108
110
  import { _GoogleMetadataRegistration } from '../llm-providers/google/register-configuration';
109
- import { joinLlmExecutionTools } from '../llm-providers/multiple/joinLlmExecutionTools';
110
- import { MultipleLlmExecutionTools } from '../llm-providers/multiple/MultipleLlmExecutionTools';
111
111
  import { _OpenAiMetadataRegistration } from '../llm-providers/openai/register-configuration';
112
112
  import { _OpenAiAssistantMetadataRegistration } from '../llm-providers/openai/register-configuration';
113
113
  import { migratePipeline } from '../migrations/migratePipeline';
@@ -249,12 +249,12 @@ export { createLlmToolsFromConfiguration };
249
249
  export { cacheLlmTools };
250
250
  export { countUsage };
251
251
  export { limitTotalUsage };
252
+ export { joinLlmExecutionTools };
253
+ export { MultipleLlmExecutionTools };
252
254
  export { _AnthropicClaudeMetadataRegistration };
253
255
  export { _AzureOpenAiMetadataRegistration };
254
256
  export { _DeepseekMetadataRegistration };
255
257
  export { _GoogleMetadataRegistration };
256
- export { joinLlmExecutionTools };
257
- export { MultipleLlmExecutionTools };
258
258
  export { _OpenAiMetadataRegistration };
259
259
  export { _OpenAiAssistantMetadataRegistration };
260
260
  export { migratePipeline };
@@ -134,6 +134,7 @@ import type { JavascriptExecutionToolsOptions } from '../scripting/javascript/Ja
134
134
  import type { PostprocessingFunction } from '../scripting/javascript/JavascriptExecutionToolsOptions';
135
135
  import type { PromptbookStorage } from '../storage/_common/PromptbookStorage';
136
136
  import type { FileCacheStorageOptions } from '../storage/file-cache-storage/FileCacheStorageOptions';
137
+ import type { IndexedDbStorageOptions } from '../storage/local-storage/utils/IndexedDbStorageOptions';
137
138
  import type { IntermediateFilesStrategy } from '../types/IntermediateFilesStrategy';
138
139
  import type { ModelRequirements } from '../types/ModelRequirements';
139
140
  import type { CompletionModelRequirements } from '../types/ModelRequirements';
@@ -431,6 +432,7 @@ export type { JavascriptExecutionToolsOptions };
431
432
  export type { PostprocessingFunction };
432
433
  export type { PromptbookStorage };
433
434
  export type { FileCacheStorageOptions };
435
+ export type { IndexedDbStorageOptions };
434
436
  export type { IntermediateFilesStrategy };
435
437
  export type { ModelRequirements };
436
438
  export type { CompletionModelRequirements };
@@ -2,6 +2,7 @@ import type { Observable } from 'rxjs';
2
2
  import { PartialDeep } from 'type-fest';
3
3
  import type { task_id } from '../types/typeAliases';
4
4
  import type { string_SCREAMING_CASE } from '../utils/normalization/normalizeTo_SCREAMING_CASE';
5
+ import type { string_promptbook_version } from '../version';
5
6
  import type { AbstractTaskResult } from './AbstractTaskResult';
6
7
  import type { PipelineExecutorResult } from './PipelineExecutorResult';
7
8
  /**
@@ -12,12 +13,21 @@ type CreateTaskOptions<TTaskResult extends AbstractTaskResult> = {
12
13
  * The type of task to create
13
14
  */
14
15
  readonly taskType: AbstractTask<TTaskResult>['taskType'];
16
+ /**
17
+ * Human-readable title of the task - used for displaying in the UI
18
+ */
19
+ readonly title: AbstractTask<TTaskResult>['title'];
15
20
  /**
16
21
  * Callback that processes the task and updates the ongoing result
17
22
  * @param ongoingResult The partial result of the task processing
18
23
  * @returns The final task result
19
24
  */
20
- taskProcessCallback(updateOngoingResult: (newOngoingResult: PartialDeep<TTaskResult>) => void): Promise<TTaskResult>;
25
+ taskProcessCallback(updateOngoingResult: (newOngoingResult: PartialDeep<TTaskResult> & {
26
+ /**
27
+ * Optional update of the task title
28
+ */
29
+ readonly title?: AbstractTask<TTaskResult>['title'];
30
+ }) => void): Promise<TTaskResult>;
21
31
  };
22
32
  /**
23
33
  * Helper to create a new task
@@ -52,10 +62,18 @@ export type AbstractTask<TTaskResult extends AbstractTaskResult> = {
52
62
  * Type of the task
53
63
  */
54
64
  readonly taskType: string_SCREAMING_CASE;
65
+ /**
66
+ * Version of the promptbook used to run the task
67
+ */
68
+ readonly promptbookVersion: string_promptbook_version;
55
69
  /**
56
70
  * Unique identifier for the task
57
71
  */
58
72
  readonly taskId: task_id;
73
+ /**
74
+ * Human-readable title of the task - used for displaying in the UI
75
+ */
76
+ readonly title: string;
59
77
  /**
60
78
  * Status of the task
61
79
  */
@@ -1,4 +1,4 @@
1
- import { MultipleLlmExecutionTools } from '../../multiple/MultipleLlmExecutionTools';
1
+ import { MultipleLlmExecutionTools } from '../../_multiple/MultipleLlmExecutionTools';
2
2
  import type { CreateLlmToolsFromConfigurationOptions } from './createLlmToolsFromConfiguration';
3
3
  /**
4
4
  * Automatically configures LLM tools from environment variables in Node.js
@@ -1,5 +1,5 @@
1
1
  import type { string_user_id } from '../../../types/typeAliases';
2
- import { MultipleLlmExecutionTools } from '../../multiple/MultipleLlmExecutionTools';
2
+ import { MultipleLlmExecutionTools } from '../../_multiple/MultipleLlmExecutionTools';
3
3
  import type { LlmToolsConfiguration } from './LlmToolsConfiguration';
4
4
  /**
5
5
  * Options for `$provideLlmToolsFromEnv`
@@ -9,6 +9,10 @@ import type { string_user_id } from '../../types/typeAliases';
9
9
  * @public exported from `@promptbook/azure-openai`
10
10
  */
11
11
  export type AzureOpenAiExecutionToolsOptions = CommonToolsOptions & {
12
+ /**
13
+ * The API key of the Azure OpenAI resource
14
+ */
15
+ readonly apiKey: string_token;
12
16
  /**
13
17
  * The resource name of the Azure OpenAI resource
14
18
  *
@@ -23,10 +27,6 @@ export type AzureOpenAiExecutionToolsOptions = CommonToolsOptions & {
23
27
  * Note: Typically you have one resource and multiple deployments.
24
28
  */
25
29
  readonly deploymentName: string_name;
26
- /**
27
- * The API key of the Azure OpenAI resource
28
- */
29
- readonly apiKey: string_token;
30
30
  /**
31
31
  * A unique identifier representing your end-user, which can help Azure OpenAI to monitor
32
32
  * and detect abuse.
@@ -1,10 +1,11 @@
1
1
  import type { PromptbookStorage } from '../_common/PromptbookStorage';
2
+ import type { IndexedDbStorageOptions } from './utils/IndexedDbStorageOptions';
2
3
  /**
3
4
  * Gets wrapper around IndexedDB which can be used as PromptbookStorage
4
5
  *
5
6
  * @public exported from `@promptbook/browser`
6
7
  */
7
- export declare function getIndexedDbStorage<TItem>(): PromptbookStorage<TItem>;
8
+ export declare function getIndexedDbStorage<TItem>(options: IndexedDbStorageOptions): PromptbookStorage<TItem>;
8
9
  /**
9
10
  * Note: [🔵] Code in this file should never be published outside of `@promptbook/browser`
10
11
  */
@@ -0,0 +1,14 @@
1
+ import type { string_name } from '../../../types/typeAliases';
2
+ /**
3
+ * Options for IndexedDB storage
4
+ */
5
+ export type IndexedDbStorageOptions = {
6
+ /**
7
+ * Name of the database
8
+ */
9
+ databaseName: string_name;
10
+ /**
11
+ * Name of the object store (table) in the database
12
+ */
13
+ storeName: string_name;
14
+ };
@@ -1,7 +1,8 @@
1
1
  import type { PromptbookStorage } from '../../_common/PromptbookStorage';
2
+ import type { IndexedDbStorageOptions } from './IndexedDbStorageOptions';
2
3
  /**
3
4
  * Creates a PromptbookStorage backed by IndexedDB.
4
5
  * Uses a single object store named 'promptbook'.
5
6
  * @private for `getIndexedDbStorage`
6
7
  */
7
- export declare function makePromptbookStorageFromIndexedDb<TValue>(dbName?: string, storeName?: string): PromptbookStorage<TValue>;
8
+ export declare function makePromptbookStorageFromIndexedDb<TValue>(options: IndexedDbStorageOptions): PromptbookStorage<TValue>;
@@ -15,7 +15,7 @@ export declare const BOOK_LANGUAGE_VERSION: string_semantic_version;
15
15
  export declare const PROMPTBOOK_ENGINE_VERSION: string_promptbook_version;
16
16
  /**
17
17
  * Represents the version string of the Promptbook engine.
18
- * It follows semantic versioning (e.g., `0.92.0-31`).
18
+ * It follows semantic versioning (e.g., `0.92.0-33`).
19
19
  *
20
20
  * @generated
21
21
  */
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@promptbook/markdown-utils",
3
- "version": "0.92.0-32",
3
+ "version": "0.92.0-34",
4
4
  "description": "It's time for a paradigm shift. The future of software in plain English, French or Latin",
5
5
  "private": false,
6
6
  "sideEffects": false,
package/umd/index.umd.js CHANGED
@@ -25,7 +25,7 @@
25
25
  * @generated
26
26
  * @see https://github.com/webgptorg/promptbook
27
27
  */
28
- const PROMPTBOOK_ENGINE_VERSION = '0.92.0-32';
28
+ const PROMPTBOOK_ENGINE_VERSION = '0.92.0-34';
29
29
  /**
30
30
  * TODO: string_promptbook_version should be constrained to the all versions of Promptbook engine
31
31
  * Note: [💞] Ignore a discrepancy between file name and entity name
@@ -2200,6 +2200,7 @@
2200
2200
  */
2201
2201
  function createTask(options) {
2202
2202
  const { taskType, taskProcessCallback } = options;
2203
+ let { title } = options;
2203
2204
  // TODO: [🐙] DRY
2204
2205
  const taskId = `${taskType.toLowerCase().substring(0, 4)}-${$randomToken(8 /* <- TODO: To global config + Use Base58 to avoid simmilar char conflicts */)}`;
2205
2206
  let status = 'RUNNING';
@@ -2211,6 +2212,10 @@
2211
2212
  const partialResultSubject = new rxjs.Subject();
2212
2213
  // <- Note: Not using `BehaviorSubject` because on error we can't access the last value
2213
2214
  const finalResultPromise = /* not await */ taskProcessCallback((newOngoingResult) => {
2215
+ if (newOngoingResult.title) {
2216
+ title = newOngoingResult.title;
2217
+ }
2218
+ updatedAt = new Date();
2214
2219
  Object.assign(currentValue, newOngoingResult);
2215
2220
  // <- TODO: assign deep
2216
2221
  partialResultSubject.next(newOngoingResult);
@@ -2256,17 +2261,24 @@
2256
2261
  return {
2257
2262
  taskType,
2258
2263
  taskId,
2264
+ get promptbookVersion() {
2265
+ return PROMPTBOOK_ENGINE_VERSION;
2266
+ },
2267
+ get title() {
2268
+ return title;
2269
+ // <- Note: [1] Theese must be getters to allow changing the value in the future
2270
+ },
2259
2271
  get status() {
2260
2272
  return status;
2261
- // <- Note: [1] Theese must be getters to allow changing the value in the future
2273
+ // <- Note: [1] --||--
2262
2274
  },
2263
2275
  get createdAt() {
2264
2276
  return createdAt;
2265
- // <- Note: [1]
2277
+ // <- Note: [1] --||--
2266
2278
  },
2267
2279
  get updatedAt() {
2268
2280
  return updatedAt;
2269
- // <- Note: [1]
2281
+ // <- Note: [1] --||--
2270
2282
  },
2271
2283
  asPromise,
2272
2284
  asObservable() {
@@ -2274,15 +2286,15 @@
2274
2286
  },
2275
2287
  get errors() {
2276
2288
  return errors;
2277
- // <- Note: [1]
2289
+ // <- Note: [1] --||--
2278
2290
  },
2279
2291
  get warnings() {
2280
2292
  return warnings;
2281
- // <- Note: [1]
2293
+ // <- Note: [1] --||--
2282
2294
  },
2283
2295
  get currentValue() {
2284
2296
  return currentValue;
2285
- // <- Note: [1]
2297
+ // <- Note: [1] --||--
2286
2298
  },
2287
2299
  };
2288
2300
  }
@@ -2447,12 +2459,14 @@
2447
2459
  const spending = new rxjs.Subject();
2448
2460
  const proxyTools = {
2449
2461
  get title() {
2450
- // TODO: [🧠] Maybe put here some suffix
2451
- return llmTools.title;
2462
+ return `${llmTools.title} (+usage)`;
2463
+ // <- TODO: [🧈] Maybe standartize the suffix when wrapping `LlmExecutionTools` up
2464
+ // <- TODO: [🧈][🧠] Does it make sence to suffix "(+usage)"?
2452
2465
  },
2453
2466
  get description() {
2454
- // TODO: [🧠] Maybe put here some suffix
2455
- return llmTools.description;
2467
+ return `${llmTools.description} (+usage)`;
2468
+ // <- TODO: [🧈] Maybe standartize the suffix when wrapping `LlmExecutionTools` up
2469
+ // <- TODO: [🧈][🧠] Does it make sence to suffix "(+usage)"?
2456
2470
  },
2457
2471
  checkConfiguration() {
2458
2472
  return /* not await */ llmTools.checkConfiguration();
@@ -2523,7 +2537,14 @@
2523
2537
  return 'Multiple LLM Providers';
2524
2538
  }
2525
2539
  get description() {
2526
- return this.llmExecutionTools.map(({ title }, index) => `${index + 1}) \`${title}\``).join('\n');
2540
+ const innerModelsTitlesAndDescriptions = this.llmExecutionTools
2541
+ .map(({ title, description }, index) => `${index + 1}) \`${title}\`\n${description}`)
2542
+ .join('\n\n');
2543
+ return spaceTrim__default["default"]((block) => `
2544
+ Multiple LLM Providers:
2545
+
2546
+ ${block(innerModelsTitlesAndDescriptions)}
2547
+ `);
2527
2548
  }
2528
2549
  /**
2529
2550
  * Check the configuration of all execution tools
@@ -5297,6 +5318,7 @@
5297
5318
  */
5298
5319
  async function getKnowledgeForTask(options) {
5299
5320
  const { tools, preparedPipeline, task, parameters } = options;
5321
+ console.log('!!! getKnowledgeForTask', options);
5300
5322
  const firstKnowlegePiece = preparedPipeline.knowledgePieces[0];
5301
5323
  const firstKnowlegeIndex = firstKnowlegePiece === null || firstKnowlegePiece === void 0 ? void 0 : firstKnowlegePiece.index[0];
5302
5324
  // <- TODO: Do not use just first knowledge piece and first index to determine embedding model, use also keyword search
@@ -5335,7 +5357,7 @@
5335
5357
  });
5336
5358
  const knowledgePiecesSorted = knowledgePiecesWithRelevance.sort((a, b) => a.relevance - b.relevance);
5337
5359
  const knowledgePiecesLimited = knowledgePiecesSorted.slice(0, 5);
5338
- console.log('!!! Embedding', {
5360
+ console.log('!!! `getKnowledgeForTask` Embedding', {
5339
5361
  task,
5340
5362
  taskEmbeddingPrompt,
5341
5363
  taskEmbeddingResult,
@@ -5371,6 +5393,7 @@
5371
5393
  */
5372
5394
  async function getReservedParametersForTask(options) {
5373
5395
  const { tools, preparedPipeline, task, parameters, pipelineIdentification } = options;
5396
+ console.log('!!! getReservedParametersForTask', options);
5374
5397
  const context = await getContextForTask(); // <- [🏍]
5375
5398
  const knowledge = await getKnowledgeForTask({ tools, preparedPipeline, task, parameters });
5376
5399
  const examples = await getExamplesForTask();
@@ -5407,6 +5430,7 @@
5407
5430
  */
5408
5431
  async function executeTask(options) {
5409
5432
  const { currentTask, preparedPipeline, parametersToPass, tools, onProgress, $executionReport, pipelineIdentification, maxExecutionAttempts, maxParallelCount, csvSettings, isVerbose, rootDirname, cacheDirname, intermediateFilesStrategy, isAutoInstalled, isNotPreparedWarningSupressed, } = options;
5433
+ console.log('!!! executeTask', options);
5410
5434
  const priority = preparedPipeline.tasks.length - preparedPipeline.tasks.indexOf(currentTask);
5411
5435
  // Note: Check consistency of used and dependent parameters which was also done in `validatePipeline`, but it’s good to doublecheck
5412
5436
  const usedParameterNames = extractParameterNamesFromTask(currentTask);
@@ -5430,14 +5454,15 @@
5430
5454
 
5431
5455
  `));
5432
5456
  }
5457
+ const reservedParameters = await getReservedParametersForTask({
5458
+ tools,
5459
+ preparedPipeline,
5460
+ task: currentTask,
5461
+ pipelineIdentification,
5462
+ parameters: parametersToPass,
5463
+ });
5433
5464
  const definedParameters = Object.freeze({
5434
- ...(await getReservedParametersForTask({
5435
- tools,
5436
- preparedPipeline,
5437
- task: currentTask,
5438
- pipelineIdentification,
5439
- parameters: parametersToPass,
5440
- })),
5465
+ ...reservedParameters,
5441
5466
  ...parametersToPass,
5442
5467
  });
5443
5468
  const definedParameterNames = new Set(Object.keys(definedParameters));
@@ -5884,6 +5909,7 @@
5884
5909
  };
5885
5910
  const pipelineExecutor = (inputParameters) => createTask({
5886
5911
  taskType: 'EXECUTION',
5912
+ title: pipeline.title,
5887
5913
  taskProcessCallback(updateOngoingResult) {
5888
5914
  return pipelineExecutorWithCallback(inputParameters, async (newOngoingResult) => {
5889
5915
  updateOngoingResult(newOngoingResult);
@@ -6014,6 +6040,12 @@
6014
6040
  }
6015
6041
  // ---
6016
6042
  if (!llmTools.callEmbeddingModel) {
6043
+ console.log('!!! No callEmbeddingModel function provided', {
6044
+ 'llmTools.title': llmTools.title,
6045
+ 'llmTools.description': llmTools.description,
6046
+ 'llmTools.callEmbeddingModel': llmTools.callEmbeddingModel,
6047
+ llmTools,
6048
+ });
6017
6049
  // TODO: [🟥] Detect browser / node and make it colorfull
6018
6050
  console.error('No callEmbeddingModel function provided');
6019
6051
  }