@promptbook/core 0.92.0-27 → 0.92.0-29

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (38) hide show
  1. package/esm/index.es.js +78 -74
  2. package/esm/index.es.js.map +1 -1
  3. package/esm/typings/src/_packages/core.index.d.ts +4 -4
  4. package/esm/typings/src/commands/FOREACH/foreachCommandParser.d.ts +0 -2
  5. package/esm/typings/src/constants.d.ts +35 -0
  6. package/esm/typings/src/executables/$provideExecutablesForNode.d.ts +1 -1
  7. package/esm/typings/src/executables/apps/locateLibreoffice.d.ts +2 -1
  8. package/esm/typings/src/executables/apps/locatePandoc.d.ts +2 -1
  9. package/esm/typings/src/executables/platforms/locateAppOnLinux.d.ts +2 -1
  10. package/esm/typings/src/executables/platforms/locateAppOnMacOs.d.ts +2 -1
  11. package/esm/typings/src/executables/platforms/locateAppOnWindows.d.ts +2 -1
  12. package/esm/typings/src/execution/AbstractTaskResult.d.ts +1 -1
  13. package/esm/typings/src/execution/LlmExecutionToolsConstructor.d.ts +2 -1
  14. package/esm/typings/src/execution/PipelineExecutorResult.d.ts +1 -1
  15. package/esm/typings/src/execution/createPipelineExecutor/$OngoingTaskResult.d.ts +12 -9
  16. package/esm/typings/src/execution/createPipelineExecutor/40-executeAttempts.d.ts +20 -14
  17. package/esm/typings/src/execution/createPipelineExecutor/filterJustOutputParameters.d.ts +7 -6
  18. package/esm/typings/src/execution/createPipelineExecutor/getContextForTask.d.ts +5 -1
  19. package/esm/typings/src/execution/createPipelineExecutor/getExamplesForTask.d.ts +1 -1
  20. package/esm/typings/src/execution/createPipelineExecutor/getKnowledgeForTask.d.ts +8 -11
  21. package/esm/typings/src/execution/translation/automatic-translate/automatic-translators/LindatAutomaticTranslator.d.ts +4 -4
  22. package/esm/typings/src/formats/csv/CsvSettings.d.ts +2 -2
  23. package/esm/typings/src/formfactors/chatbot/ChatbotFormfactorDefinition.d.ts +2 -2
  24. package/esm/typings/src/formfactors/completion/CompletionFormfactorDefinition.d.ts +1 -1
  25. package/esm/typings/src/formfactors/generator/GeneratorFormfactorDefinition.d.ts +2 -1
  26. package/esm/typings/src/formfactors/generic/GenericFormfactorDefinition.d.ts +2 -2
  27. package/esm/typings/src/formfactors/index.d.ts +3 -3
  28. package/esm/typings/src/llm-providers/_common/register/LlmToolsMetadata.d.ts +4 -37
  29. package/esm/typings/src/llm-providers/anthropic-claude/anthropic-claude-models.d.ts +1 -1
  30. package/esm/typings/src/llm-providers/deepseek/deepseek-models.d.ts +1 -1
  31. package/esm/typings/src/llm-providers/google/google-models.d.ts +1 -1
  32. package/esm/typings/src/llm-providers/openai/openai-models.d.ts +1 -1
  33. package/esm/typings/src/llm-providers/openai/register-configuration.d.ts +2 -2
  34. package/esm/typings/src/llm-providers/openai/register-constructor.d.ts +2 -2
  35. package/esm/typings/src/version.d.ts +1 -1
  36. package/package.json +1 -1
  37. package/umd/index.umd.js +79 -75
  38. package/umd/index.umd.js.map +1 -1
package/esm/index.es.js CHANGED
@@ -27,7 +27,7 @@ const BOOK_LANGUAGE_VERSION = '1.0.0';
27
27
  * @generated
28
28
  * @see https://github.com/webgptorg/promptbook
29
29
  */
30
- const PROMPTBOOK_ENGINE_VERSION = '0.92.0-27';
30
+ const PROMPTBOOK_ENGINE_VERSION = '0.92.0-29';
31
31
  /**
32
32
  * TODO: string_promptbook_version should be constrained to the all versions of Promptbook engine
33
33
  * Note: [💞] Ignore a discrepancy between file name and entity name
@@ -1099,6 +1099,42 @@ function exportJson(options) {
1099
1099
  * TODO: [🧠] Is there a way how to meaningfully test this utility
1100
1100
  */
1101
1101
 
1102
+ /**
1103
+ * How is the model provider trusted?
1104
+ *
1105
+ * @public exported from `@promptbook/core`
1106
+ */
1107
+ const MODEL_TRUST_LEVELS = {
1108
+ FULL: `Model is running on the local machine, training data and model weights are known, data are ethically sourced`,
1109
+ OPEN: `Model is open source, training data and model weights are known`,
1110
+ PARTIALLY_OPEN: `Model is open source, but training data and model weights are not (fully) known`,
1111
+ CLOSED_LOCAL: `Model can be run locally, but it is not open source`,
1112
+ CLOSED_FREE: `Model is behind API gateway but free to use`,
1113
+ CLOSED_BUSINESS: `Model is behind API gateway and paid but has good SLA, TOS, privacy policy and in general is a good to use in business applications`,
1114
+ CLOSED: `Model is behind API gateway and paid`,
1115
+ UNTRUSTED: `Model has questions about the training data and ethics, but it is not known if it is a problem or not`,
1116
+ VURNABLE: `Model has some known serious vulnerabilities, leaks, ethical problems, etc.`,
1117
+ };
1118
+ // <- TODO: Maybe do better levels of trust
1119
+ /**
1120
+ * How is the model provider important?
1121
+ *
1122
+ * @public exported from `@promptbook/core`
1123
+ */
1124
+ const MODEL_ORDERS = {
1125
+ /**
1126
+ * Top-tier models, e.g. OpenAI, Anthropic,...
1127
+ */
1128
+ TOP_TIER: 333,
1129
+ /**
1130
+ * Mid-tier models, e.g. Llama, Mistral, etc.
1131
+ */
1132
+ NORMAL: 100,
1133
+ /**
1134
+ * Low-tier models, e.g. Phi, Tiny, etc.
1135
+ */
1136
+ LOW_TIER: 0,
1137
+ };
1102
1138
  /**
1103
1139
  * Order of keys in the pipeline JSON
1104
1140
  *
@@ -2780,7 +2816,7 @@ function union(...sets) {
2780
2816
  }
2781
2817
 
2782
2818
  /**
2783
- * @@@
2819
+ * Contains configuration options for parsing and generating CSV files, such as delimiters and quoting rules.
2784
2820
  *
2785
2821
  * @public exported from `@promptbook/core`
2786
2822
  */
@@ -3989,8 +4025,12 @@ function isPassingExpectations(expectations, value) {
3989
4025
  */
3990
4026
 
3991
4027
  /**
3992
- * @@@
4028
+ * Executes a pipeline task with multiple attempts, including joker and retry logic. Handles different task types
4029
+ * (prompt, script, dialog, etc.), applies postprocessing, checks expectations, and updates the execution report.
4030
+ * Throws errors if execution fails after all attempts.
3993
4031
  *
4032
+ * @param options - The options for execution, including task, parameters, pipeline, and configuration.
4033
+ * @returns The result string of the executed task.
3994
4034
  * @private internal utility of `createPipelineExecutor`
3995
4035
  */
3996
4036
  async function executeAttempts(options) {
@@ -4448,8 +4488,12 @@ async function executeFormatSubvalues(options) {
4448
4488
  }
4449
4489
 
4450
4490
  /**
4451
- * @@@
4491
+ * Returns the context for a given task, typically used to provide additional information or variables
4492
+ * required for the execution of the task within a pipeline. The context is returned as a string value
4493
+ * that may include markdown formatting.
4452
4494
  *
4495
+ * @param task - The task for which the context is being generated. This should be a deeply immutable TaskJson object.
4496
+ * @returns The context as a string, formatted as markdown and parameter value.
4453
4497
  * @private internal utility of `createPipelineExecutor`
4454
4498
  */
4455
4499
  async function getContextForTask(task) {
@@ -4457,7 +4501,7 @@ async function getContextForTask(task) {
4457
4501
  }
4458
4502
 
4459
4503
  /**
4460
- * @@@
4504
+ * Retrieves example values or templates for a given task, used to guide or validate pipeline execution.
4461
4505
  *
4462
4506
  * @private internal utility of `createPipelineExecutor`
4463
4507
  */
@@ -4504,9 +4548,8 @@ function knowledgePiecesToString(knowledgePieces) {
4504
4548
  }
4505
4549
 
4506
4550
  /**
4507
- * @@@
4508
- *
4509
- * Here is the place where RAG (retrieval-augmented generation) happens
4551
+ * Retrieves the most relevant knowledge pieces for a given task using embedding-based similarity search.
4552
+ * This is where retrieval-augmented generation (RAG) is performed to enhance the task with external knowledge.
4510
4553
  *
4511
4554
  * @private internal utility of `createPipelineExecutor`
4512
4555
  */
@@ -4725,7 +4768,8 @@ async function executeTask(options) {
4725
4768
  */
4726
4769
 
4727
4770
  /**
4728
- * @@@
4771
+ * Filters and returns only the output parameters from the provided pipeline execution options.
4772
+ * Adds warnings for any expected output parameters that are missing.
4729
4773
  *
4730
4774
  * @private internal utility of `createPipelineExecutor`
4731
4775
  */
@@ -7035,8 +7079,6 @@ function validateParameterName(parameterName) {
7035
7079
  /**
7036
7080
  * Parses the foreach command
7037
7081
  *
7038
- * Note: @@@ This command is used as foreach for new commands - it should NOT be used in any `.book` file
7039
- *
7040
7082
  * @see `documentationUrl` for more details
7041
7083
  * @public exported from `@promptbook/editable`
7042
7084
  */
@@ -7277,14 +7319,14 @@ const formatCommandParser = {
7277
7319
  };
7278
7320
 
7279
7321
  /**
7280
- * @@@
7322
+ * Chatbot form factor definition for conversational interfaces that interact with users in a chat-like manner.
7281
7323
  *
7282
7324
  * @public exported from `@promptbook/core`
7283
7325
  */
7284
7326
  const ChatbotFormfactorDefinition = {
7285
7327
  name: 'CHATBOT',
7286
7328
  aliasNames: ['CHAT'],
7287
- description: `@@@`,
7329
+ description: `A chatbot form factor for conversational user interfaces.`,
7288
7330
  documentationUrl: `https://github.com/webgptorg/promptbook/discussions/174`,
7289
7331
  pipelineInterface: {
7290
7332
  inputParameters: [
@@ -7317,7 +7359,7 @@ const ChatbotFormfactorDefinition = {
7317
7359
  */
7318
7360
  const CompletionFormfactorDefinition = {
7319
7361
  name: 'COMPLETION',
7320
- description: `@@@`,
7362
+ description: `Completion is formfactor that emulates completion models`,
7321
7363
  documentationUrl: `https://github.com/webgptorg/promptbook/discussions/@@`,
7322
7364
  // <- TODO: https://github.com/webgptorg/promptbook/discussions/new?category=concepts
7323
7365
  // "🔠 Completion Formfactor"
@@ -7348,7 +7390,8 @@ const CompletionFormfactorDefinition = {
7348
7390
  };
7349
7391
 
7350
7392
  /**
7351
- * Generator is form of app that @@@
7393
+ * Generator form factor represents an application that generates content or data based on user input or predefined rules.
7394
+ * This form factor is used for apps that produce outputs, such as text, images, or other media, based on provided input.
7352
7395
  *
7353
7396
  * @public exported from `@promptbook/core`
7354
7397
  */
@@ -7392,13 +7435,13 @@ const GENERIC_PIPELINE_INTERFACE = {
7392
7435
  */
7393
7436
 
7394
7437
  /**
7395
- * @@@
7438
+ * A generic pipeline
7396
7439
  *
7397
7440
  * @public exported from `@promptbook/core`
7398
7441
  */
7399
7442
  const GenericFormfactorDefinition = {
7400
7443
  name: 'GENERIC',
7401
- description: `@@@`,
7444
+ description: `A generic pipeline`,
7402
7445
  documentationUrl: `https://github.com/webgptorg/promptbook/discussions/173`,
7403
7446
  pipelineInterface: GENERIC_PIPELINE_INTERFACE,
7404
7447
  };
@@ -9734,13 +9777,16 @@ function addAutoGeneratedSection(content, options) {
9734
9777
  */
9735
9778
  function renderPromptbookMermaid(pipelineJson, options) {
9736
9779
  const { linkTask = () => null } = options || {};
9780
+ const MERMAID_PREFIX = 'mermaid_';
9781
+ const KNOWLEDGE_NAME = MERMAID_PREFIX + 'knowledge';
9782
+ const RESERVED_NAME = MERMAID_PREFIX + 'reserved';
9737
9783
  const parameterNameToTaskName = (parameterName) => {
9738
9784
  if (parameterName === 'knowledge') {
9739
- return 'knowledge';
9785
+ return KNOWLEDGE_NAME;
9740
9786
  // <- TODO: !!!! Check that this works
9741
9787
  }
9742
9788
  else if (RESERVED_PARAMETER_NAMES.includes(parameterName)) {
9743
- return 'reserved';
9789
+ return RESERVED_NAME;
9744
9790
  }
9745
9791
  const parameter = pipelineJson.parameters.find((parameter) => parameter.name === parameterName);
9746
9792
  if (!parameter) {
@@ -9748,13 +9794,13 @@ function renderPromptbookMermaid(pipelineJson, options) {
9748
9794
  // <- TODO: !!6 This causes problems when {knowledge} and other reserved parameters are used
9749
9795
  }
9750
9796
  if (parameter.isInput) {
9751
- return 'input';
9797
+ return MERMAID_PREFIX + 'input';
9752
9798
  }
9753
9799
  const task = pipelineJson.tasks.find((task) => task.resultingParameterName === parameterName);
9754
9800
  if (!task) {
9755
9801
  throw new Error(`Could not find task for {${parameterName}}`);
9756
9802
  }
9757
- return task.name || normalizeTo_camelCase('task-' + titleToName(task.title));
9803
+ return MERMAID_PREFIX + (task.name || normalizeTo_camelCase('task-' + titleToName(task.title)));
9758
9804
  };
9759
9805
  const promptbookMermaid = spaceTrim$1((block) => `
9760
9806
 
@@ -9766,8 +9812,8 @@ function renderPromptbookMermaid(pipelineJson, options) {
9766
9812
  direction TB
9767
9813
 
9768
9814
  input((Input)):::input
9769
- other((Other)):::other
9770
- knowledge((Knowledgebase)):::knowledge
9815
+ ${RESERVED_NAME}((Other)):::${RESERVED_NAME}
9816
+ ${KNOWLEDGE_NAME}((Knowledgebase)):::${KNOWLEDGE_NAME}
9771
9817
  ${block(pipelineJson.tasks
9772
9818
  .flatMap(({ title, dependentParameterNames, resultingParameterName }) => [
9773
9819
  `${parameterNameToTaskName(resultingParameterName)}("${title}")`,
@@ -10565,48 +10611,6 @@ function createLlmToolsFromConfiguration(configuration, options = {}) {
10565
10611
  * TODO: [®] DRY Register logic
10566
10612
  */
10567
10613
 
10568
- /**
10569
- * How is the model provider trusted?
10570
- *
10571
- * @public exported from `@promptbook/core`
10572
- */
10573
- const MODEL_TRUST_LEVEL = {
10574
- FULL: `Model is running on the local machine, training data and model weights are known, data are ethically sourced`,
10575
- OPEN: `Model is open source, training data and model weights are known`,
10576
- PARTIALLY_OPEN: `Model is open source, but training data and model weights are not (fully) known`,
10577
- CLOSED_LOCAL: `Model can be run locally, but it is not open source`,
10578
- CLOSED_FREE: `Model is behind API gateway but free to use`,
10579
- CLOSED_BUSINESS: `Model is behind API gateway and paid but has good SLA, TOS, privacy policy and in general is a good to use in business applications`,
10580
- CLOSED: `Model is behind API gateway and paid`,
10581
- UNTRUSTED: `Model has questions about the training data and ethics, but it is not known if it is a problem or not`,
10582
- VURNABLE: `Model has some known serious vulnerabilities, leaks, ethical problems, etc.`,
10583
- };
10584
- // <- TODO: Maybe do better levels of trust
10585
- /**
10586
- * How is the model provider important?
10587
- *
10588
- * @public exported from `@promptbook/core`
10589
- */
10590
- const MODEL_ORDER = {
10591
- /**
10592
- * Top-tier models, e.g. OpenAI, Anthropic,...
10593
- */
10594
- TOP_TIER: 333,
10595
- /**
10596
- * Mid-tier models, e.g. Llama, Mistral, etc.
10597
- */
10598
- NORMAL: 100,
10599
- /**
10600
- * Low-tier models, e.g. Phi, Tiny, etc.
10601
- */
10602
- LOW_TIER: 0,
10603
- };
10604
- /**
10605
- * TODO: Add configuration schema and maybe some documentation link
10606
- * TODO: Maybe constrain LlmToolsConfiguration[number] by generic to ensure that `createConfigurationFromEnv` and `getBoilerplateConfiguration` always create same `packageName` and `className`
10607
- * TODO: [®] DRY Register logic
10608
- */
10609
-
10610
10614
  /**
10611
10615
  * Stores data in memory (HEAP)
10612
10616
  *
@@ -10839,7 +10843,7 @@ const _AnthropicClaudeMetadataRegistration = $llmToolsMetadataRegister.register(
10839
10843
  className: 'AnthropicClaudeExecutionTools',
10840
10844
  envVariables: ['ANTHROPIC_CLAUDE_API_KEY'],
10841
10845
  trustLevel: 'CLOSED',
10842
- order: MODEL_ORDER.TOP_TIER,
10846
+ order: MODEL_ORDERS.TOP_TIER,
10843
10847
  getBoilerplateConfiguration() {
10844
10848
  return {
10845
10849
  title: 'Anthropic Claude',
@@ -10886,7 +10890,7 @@ const _AzureOpenAiMetadataRegistration = $llmToolsMetadataRegister.register({
10886
10890
  className: 'AzureOpenAiExecutionTools',
10887
10891
  envVariables: ['AZUREOPENAI_RESOURCE_NAME', 'AZUREOPENAI_DEPLOYMENT_NAME', 'AZUREOPENAI_API_KEY'],
10888
10892
  trustLevel: 'CLOSED_BUSINESS',
10889
- order: MODEL_ORDER.NORMAL,
10893
+ order: MODEL_ORDERS.NORMAL,
10890
10894
  getBoilerplateConfiguration() {
10891
10895
  return {
10892
10896
  title: 'Azure Open AI',
@@ -10974,7 +10978,7 @@ const _DeepseekMetadataRegistration = $llmToolsMetadataRegister.register({
10974
10978
  className: 'DeepseekExecutionTools',
10975
10979
  envVariables: ['DEEPSEEK_GENERATIVE_AI_API_KEY'],
10976
10980
  trustLevel: 'UNTRUSTED',
10977
- order: MODEL_ORDER.NORMAL,
10981
+ order: MODEL_ORDERS.NORMAL,
10978
10982
  getBoilerplateConfiguration() {
10979
10983
  return {
10980
10984
  title: 'Deepseek',
@@ -11025,7 +11029,7 @@ const _GoogleMetadataRegistration = $llmToolsMetadataRegister.register({
11025
11029
  className: 'GoogleExecutionTools',
11026
11030
  envVariables: ['GOOGLE_GENERATIVE_AI_API_KEY'],
11027
11031
  trustLevel: 'CLOSED',
11028
- order: MODEL_ORDER.NORMAL,
11032
+ order: MODEL_ORDERS.NORMAL,
11029
11033
  getBoilerplateConfiguration() {
11030
11034
  return {
11031
11035
  title: 'Google Gemini',
@@ -11076,7 +11080,7 @@ const _OpenAiMetadataRegistration = $llmToolsMetadataRegister.register({
11076
11080
  className: 'OpenAiExecutionTools',
11077
11081
  envVariables: ['OPENAI_API_KEY'],
11078
11082
  trustLevel: 'CLOSED',
11079
- order: MODEL_ORDER.TOP_TIER,
11083
+ order: MODEL_ORDERS.TOP_TIER,
11080
11084
  getBoilerplateConfiguration() {
11081
11085
  return {
11082
11086
  title: 'Open AI',
@@ -11104,9 +11108,9 @@ const _OpenAiMetadataRegistration = $llmToolsMetadataRegister.register({
11104
11108
  },
11105
11109
  });
11106
11110
  /**
11107
- * @@@ registration1 of default configuration for Open AI
11111
+ * Registration of the OpenAI Assistant metadata
11108
11112
  *
11109
- * Note: [🏐] Configurations registrations are done in @@@ BUT constructor @@@
11113
+ * Note: [🏐] Configurations registrations are done in the metadata registration section, but the constructor registration is handled separately.
11110
11114
  *
11111
11115
  * @public exported from `@promptbook/core`
11112
11116
  * @public exported from `@promptbook/wizzard`
@@ -11119,7 +11123,7 @@ const _OpenAiAssistantMetadataRegistration = $llmToolsMetadataRegister.register(
11119
11123
  envVariables: null,
11120
11124
  // <- TODO: ['OPENAI_API_KEY', 'OPENAI_ASSISTANT_ID']
11121
11125
  trustLevel: 'CLOSED',
11122
- order: MODEL_ORDER.NORMAL,
11126
+ order: MODEL_ORDERS.NORMAL,
11123
11127
  getBoilerplateConfiguration() {
11124
11128
  return {
11125
11129
  title: 'Open AI Assistant',
@@ -11632,5 +11636,5 @@ class PrefixStorage {
11632
11636
  }
11633
11637
  }
11634
11638
 
11635
- export { $llmToolsMetadataRegister, $llmToolsRegister, $scrapersMetadataRegister, $scrapersRegister, ADMIN_EMAIL, ADMIN_GITHUB_NAME, AbstractFormatError, AuthenticationError, BIG_DATASET_TRESHOLD, BOOK_LANGUAGE_VERSION, BlackholeStorage, BoilerplateError, BoilerplateFormfactorDefinition, CLAIM, CLI_APP_ID, CallbackInterfaceTools, ChatbotFormfactorDefinition, CollectionError, CompletionFormfactorDefinition, CsvFormatError, CsvFormatParser, DEFAULT_BOOKS_DIRNAME, DEFAULT_BOOK_OUTPUT_PARAMETER_NAME, DEFAULT_BOOK_TITLE, DEFAULT_CSV_SETTINGS, DEFAULT_DOWNLOAD_CACHE_DIRNAME, DEFAULT_EXECUTION_CACHE_DIRNAME, DEFAULT_GET_PIPELINE_COLLECTION_FUNCTION_NAME, DEFAULT_INTERMEDIATE_FILES_STRATEGY, DEFAULT_IS_AUTO_INSTALLED, DEFAULT_IS_VERBOSE, DEFAULT_MAX_EXECUTION_ATTEMPTS, DEFAULT_MAX_FILE_SIZE, DEFAULT_MAX_KNOWLEDGE_SOURCES_SCRAPING_DEPTH, DEFAULT_MAX_KNOWLEDGE_SOURCES_SCRAPING_TOTAL, DEFAULT_MAX_PARALLEL_COUNT, DEFAULT_PIPELINE_COLLECTION_BASE_FILENAME, DEFAULT_PROMPT_TASK_TITLE, DEFAULT_REMOTE_SERVER_URL, DEFAULT_RPM, DEFAULT_SCRAPE_CACHE_DIRNAME, DEFAULT_TASK_TITLE, EXPECTATION_UNITS, EnvironmentMismatchError, ExecutionReportStringOptionsDefaults, ExpectError, FAILED_VALUE_PLACEHOLDER, FORMFACTOR_DEFINITIONS, GENERIC_PIPELINE_INTERFACE, GeneratorFormfactorDefinition, GenericFormfactorDefinition, ImageGeneratorFormfactorDefinition, KnowledgeScrapeError, LimitReachedError, MANDATORY_CSV_SETTINGS, MAX_FILENAME_LENGTH, MODEL_ORDER, MODEL_TRUST_LEVEL, MODEL_VARIANTS, MatcherFormfactorDefinition, MemoryStorage, MissingToolsError, MultipleLlmExecutionTools, NAME, NonTaskSectionTypes, NotFoundError, NotYetImplementedError, ORDER_OF_PIPELINE_JSON, PENDING_VALUE_PLACEHOLDER, PLAYGROUND_APP_ID, PROMPTBOOK_ENGINE_VERSION, PROMPTBOOK_ERRORS, ParseError, PipelineExecutionError, PipelineLogicError, PipelineUrlError, PrefixStorage, PromptbookFetchError, REMOTE_SERVER_URLS, RESERVED_PARAMETER_NAMES, SET_IS_VERBOSE, SectionTypes, SheetsFormfactorDefinition, TaskTypes, TextFormatParser, TranslatorFormfactorDefinition, UNCERTAIN_USAGE, UNCERTAIN_ZERO_VALUE, UnexpectedError, WrappedError, ZERO_USAGE, ZERO_VALUE, _AnthropicClaudeMetadataRegistration, _AzureOpenAiMetadataRegistration, _BoilerplateScraperMetadataRegistration, _DeepseekMetadataRegistration, _DocumentScraperMetadataRegistration, _GoogleMetadataRegistration, _LegacyDocumentScraperMetadataRegistration, _MarkdownScraperMetadataRegistration, _MarkitdownScraperMetadataRegistration, _OpenAiAssistantMetadataRegistration, _OpenAiMetadataRegistration, _PdfScraperMetadataRegistration, _WebsiteScraperMetadataRegistration, addUsage, book, cacheLlmTools, collectionToJson, compilePipeline, computeCosineSimilarity, countUsage, createCollectionFromJson, createCollectionFromPromise, createCollectionFromUrl, createLlmToolsFromConfiguration, createPipelineExecutor, createSubcollection, embeddingVectorToString, executionReportJsonToString, extractParameterNamesFromTask, filterModels, getPipelineInterface, identificationToPromptbookToken, isPassingExpectations, isPipelineImplementingInterface, isPipelineInterfacesEqual, isPipelinePrepared, isValidPipelineString, joinLlmExecutionTools, limitTotalUsage, makeKnowledgeSourceHandler, migratePipeline, parsePipeline, pipelineJsonToString, prepareKnowledgePieces, preparePersona, preparePipeline, prettifyPipelineString, promptbookFetch, promptbookTokenToIdentification, unpreparePipeline, usageToHuman, usageToWorktime, validatePipeline, validatePipelineString };
11639
+ export { $llmToolsMetadataRegister, $llmToolsRegister, $scrapersMetadataRegister, $scrapersRegister, ADMIN_EMAIL, ADMIN_GITHUB_NAME, AbstractFormatError, AuthenticationError, BIG_DATASET_TRESHOLD, BOOK_LANGUAGE_VERSION, BlackholeStorage, BoilerplateError, BoilerplateFormfactorDefinition, CLAIM, CLI_APP_ID, CallbackInterfaceTools, ChatbotFormfactorDefinition, CollectionError, CompletionFormfactorDefinition, CsvFormatError, CsvFormatParser, DEFAULT_BOOKS_DIRNAME, DEFAULT_BOOK_OUTPUT_PARAMETER_NAME, DEFAULT_BOOK_TITLE, DEFAULT_CSV_SETTINGS, DEFAULT_DOWNLOAD_CACHE_DIRNAME, DEFAULT_EXECUTION_CACHE_DIRNAME, DEFAULT_GET_PIPELINE_COLLECTION_FUNCTION_NAME, DEFAULT_INTERMEDIATE_FILES_STRATEGY, DEFAULT_IS_AUTO_INSTALLED, DEFAULT_IS_VERBOSE, DEFAULT_MAX_EXECUTION_ATTEMPTS, DEFAULT_MAX_FILE_SIZE, DEFAULT_MAX_KNOWLEDGE_SOURCES_SCRAPING_DEPTH, DEFAULT_MAX_KNOWLEDGE_SOURCES_SCRAPING_TOTAL, DEFAULT_MAX_PARALLEL_COUNT, DEFAULT_PIPELINE_COLLECTION_BASE_FILENAME, DEFAULT_PROMPT_TASK_TITLE, DEFAULT_REMOTE_SERVER_URL, DEFAULT_RPM, DEFAULT_SCRAPE_CACHE_DIRNAME, DEFAULT_TASK_TITLE, EXPECTATION_UNITS, EnvironmentMismatchError, ExecutionReportStringOptionsDefaults, ExpectError, FAILED_VALUE_PLACEHOLDER, FORMFACTOR_DEFINITIONS, GENERIC_PIPELINE_INTERFACE, GeneratorFormfactorDefinition, GenericFormfactorDefinition, ImageGeneratorFormfactorDefinition, KnowledgeScrapeError, LimitReachedError, MANDATORY_CSV_SETTINGS, MAX_FILENAME_LENGTH, MODEL_ORDERS, MODEL_TRUST_LEVELS, MODEL_VARIANTS, MatcherFormfactorDefinition, MemoryStorage, MissingToolsError, MultipleLlmExecutionTools, NAME, NonTaskSectionTypes, NotFoundError, NotYetImplementedError, ORDER_OF_PIPELINE_JSON, PENDING_VALUE_PLACEHOLDER, PLAYGROUND_APP_ID, PROMPTBOOK_ENGINE_VERSION, PROMPTBOOK_ERRORS, ParseError, PipelineExecutionError, PipelineLogicError, PipelineUrlError, PrefixStorage, PromptbookFetchError, REMOTE_SERVER_URLS, RESERVED_PARAMETER_NAMES, SET_IS_VERBOSE, SectionTypes, SheetsFormfactorDefinition, TaskTypes, TextFormatParser, TranslatorFormfactorDefinition, UNCERTAIN_USAGE, UNCERTAIN_ZERO_VALUE, UnexpectedError, WrappedError, ZERO_USAGE, ZERO_VALUE, _AnthropicClaudeMetadataRegistration, _AzureOpenAiMetadataRegistration, _BoilerplateScraperMetadataRegistration, _DeepseekMetadataRegistration, _DocumentScraperMetadataRegistration, _GoogleMetadataRegistration, _LegacyDocumentScraperMetadataRegistration, _MarkdownScraperMetadataRegistration, _MarkitdownScraperMetadataRegistration, _OpenAiAssistantMetadataRegistration, _OpenAiMetadataRegistration, _PdfScraperMetadataRegistration, _WebsiteScraperMetadataRegistration, addUsage, book, cacheLlmTools, collectionToJson, compilePipeline, computeCosineSimilarity, countUsage, createCollectionFromJson, createCollectionFromPromise, createCollectionFromUrl, createLlmToolsFromConfiguration, createPipelineExecutor, createSubcollection, embeddingVectorToString, executionReportJsonToString, extractParameterNamesFromTask, filterModels, getPipelineInterface, identificationToPromptbookToken, isPassingExpectations, isPipelineImplementingInterface, isPipelineInterfacesEqual, isPipelinePrepared, isValidPipelineString, joinLlmExecutionTools, limitTotalUsage, makeKnowledgeSourceHandler, migratePipeline, parsePipeline, pipelineJsonToString, prepareKnowledgePieces, preparePersona, preparePipeline, prettifyPipelineString, promptbookFetch, promptbookTokenToIdentification, unpreparePipeline, usageToHuman, usageToWorktime, validatePipeline, validatePipelineString };
11636
11640
  //# sourceMappingURL=index.es.js.map