@promptbook/node 0.72.0 → 0.73.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -12,4 +12,4 @@ export type TemplateType = TupleToUnion<typeof TemplateTypes>;
12
12
  * @see https://github.com/webgptorg/promptbook#template-type
13
13
  * @public exported from `@promptbook/core`
14
14
  */
15
- export declare const TemplateTypes: readonly ["PROMPT_TEMPLATE", "SIMPLE_TEMPLATE", "SCRIPT_TEMPLATE", "DIALOG_TEMPLATE", "SAMPLE", "KNOWLEDGE", "INSTRUMENT", "ACTION"];
15
+ export declare const TemplateTypes: readonly ["PROMPT_TEMPLATE", "SIMPLE_TEMPLATE", "SCRIPT_TEMPLATE", "DIALOG_TEMPLATE", "EXAMPLE", "KNOWLEDGE", "INSTRUMENT", "ACTION"];
@@ -121,7 +121,7 @@ export declare const REPLACING_NONCE = "u$k42k%!V2zo34w7Fu#@QUHYPW";
121
121
  *
122
122
  * @public exported from `@promptbook/core`
123
123
  */
124
- export declare const RESERVED_PARAMETER_NAMES: readonly ["content", "context", "knowledge", "samples", "modelName", "currentDate"];
124
+ export declare const RESERVED_PARAMETER_NAMES: readonly ["content", "context", "knowledge", "examples", "modelName", "currentDate"];
125
125
  /**
126
126
  * @@@
127
127
  *
@@ -22,9 +22,9 @@ export type renderPipelineMermaidOptions = {
22
22
  */
23
23
  export declare function renderPromptbookMermaid(pipelineJson: PipelineJson, options?: renderPipelineMermaidOptions): string;
24
24
  /**
25
- * TODO: !!!!! FOREACH in mermaid graph
26
- * TODO: !!!!! Knowledge in mermaid graph
27
- * TODO: !!!!! Personas in mermaid graph
25
+ * TODO: [🧠] !! FOREACH in mermaid graph
26
+ * TODO: [🧠] !! Knowledge in mermaid graph
27
+ * TODO: [🧠] !! Personas in mermaid graph
28
28
  * TODO: Maybe use some Mermaid package instead of string templating
29
29
  * TODO: [🕌] When more than 2 functionalities, split into separate functions
30
30
  */
@@ -9,7 +9,7 @@ import type { string_json } from '../../types/typeAliases';
9
9
  */
10
10
  export declare function stringifyPipelineJson<TType>(pipeline: TType): string_json<TType>;
11
11
  /**
12
- * TODO: [🐝] Not Working propperly @see https://promptbook.studio/samples/mixed-knowledge.ptbk.md
12
+ * TODO: [🐝] Not Working propperly @see https://promptbook.studio/examples/mixed-knowledge.ptbk.md
13
13
  * TODO: [🧠][0] Maybe rename to `stringifyPipelineJson`, `stringifyIndexedJson`,...
14
14
  * TODO: [🧠] Maybe more elegant solution than replacing via regex
15
15
  * TODO: [🍙] Make some standard order of json properties
@@ -7,7 +7,7 @@ import type { string_json } from '../../types/typeAliases';
7
7
  * Note: Using here custom import to work in jest tests
8
8
  * Note: Using sync version is 💩 in the production code, but it's ok here in tests
9
9
  *
10
- * @param path - The path to the file relative to samples/pipelines directory
10
+ * @param path - The path to the file relative to examples/pipelines directory
11
11
  * @private internal function of tests
12
12
  */
13
13
  export declare function importPipelineWithoutPreparation(path: `${string}.ptbk.md`): PipelineString;
@@ -20,7 +20,7 @@ export declare function validatePipeline(pipeline: PipelineJson): PipelineJson;
20
20
  */
21
21
  export declare function validatePipelineCore(pipeline: PipelineJson): void;
22
22
  /**
23
- * TODO: !!!!! [🧞‍♀️] Do not allow joker + foreach
23
+ * TODO: !! [🧞‍♀️] Do not allow joker + foreach
24
24
  * TODO: [🧠] Work with promptbookVersion
25
25
  * TODO: Use here some json-schema, Zod or something similar and change it to:
26
26
  * > /**
@@ -32,7 +32,7 @@ export declare function validatePipelineCore(pipeline: PipelineJson): void;
32
32
  * > ex port function validatePipeline(promptbook: really_unknown): asserts promptbook is PipelineJson {
33
33
  */
34
34
  /**
35
- * TODO: [🧳][main] !!!! Validate that all samples match expectations
35
+ * TODO: [🧳][main] !!!! Validate that all examples match expectations
36
36
  * TODO: [🧳][🐝][main] !!!! Validate that knowledge is valid (non-void)
37
37
  * TODO: [🧳][main] !!!! Validate that persona can be used only with CHAT variant
38
38
  * TODO: [🧳][main] !!!! Validate that parameter with reserved name not used RESERVED_PARAMETER_NAMES
@@ -7,4 +7,4 @@ import type { string_parameter_value } from '../../types/typeAliases';
7
7
  *
8
8
  * @private internal utility of `createPipelineExecutor`
9
9
  */
10
- export declare function getSamplesForTemplate(template: ReadonlyDeep<TemplateJson>): Promise<string_parameter_value & string_markdown>;
10
+ export declare function getExamplesForTemplate(template: ReadonlyDeep<TemplateJson>): Promise<string_parameter_value & string_markdown>;
@@ -15,7 +15,7 @@ export type FormatDefinition<TValue extends TPartialValue, TPartialValue extends
15
15
  /**
16
16
  * The name of the format used in .ptbk.md files
17
17
  *
18
- * @sample "JSON"
18
+ * @example "JSON"
19
19
  */
20
20
  readonly formatName: string_name & string_SCREAMING_CASE;
21
21
  /**
@@ -25,7 +25,7 @@ export type FormatDefinition<TValue extends TPartialValue, TPartialValue extends
25
25
  /**
26
26
  * The mime type of the format (if any)
27
27
  *
28
- * @sample "application/json"
28
+ * @example "application/json"
29
29
  */
30
30
  readonly mimeType?: string_mime_type;
31
31
  /**
@@ -11,7 +11,7 @@ export type FormatSubvalueDefinition<TValue extends string, TSettings extends em
11
11
  /**
12
12
  * The name of the format used in .ptbk.md files
13
13
  *
14
- * @sample "CELL"
14
+ * @example "CELL"
15
15
  */
16
16
  readonly subvalueName: string_name & string_SCREAMING_CASE;
17
17
  /**
@@ -10,7 +10,7 @@ import type { string_persona_description } from '../types/typeAliases';
10
10
  */
11
11
  export declare function preparePersona(personaDescription: string_persona_description, tools: Pick<ExecutionTools, 'llm'>, options: PrepareAndScrapeOptions): Promise<PersonaPreparedJson['modelRequirements']>;
12
12
  /**
13
- * TODO: [🔃][main] !!!!! If the persona was prepared with different version or different set of models, prepare it once again
13
+ * TODO: [🔃][main] !! If the persona was prepared with different version or different set of models, prepare it once again
14
14
  * TODO: [🏢] !! Check validity of `modelName` in pipeline
15
15
  * TODO: [🏢] !! Check validity of `systemMessage` in pipeline
16
16
  * TODO: [🏢] !! Check validity of `temperature` in pipeline
@@ -6,11 +6,11 @@ import type { PipelineJson } from '../types/PipelineJson/PipelineJson';
6
6
  */
7
7
  export declare function isPipelinePrepared(pipeline: PipelineJson): boolean;
8
8
  /**
9
- * TODO: [🔃][main] !!!!! If the pipeline was prepared with different version or different set of models, prepare it once again
9
+ * TODO: [🔃][main] !! If the pipeline was prepared with different version or different set of models, prepare it once again
10
10
  * TODO: [🐠] Maybe base this on `makeValidator`
11
11
  * TODO: [🧊] Pipeline can be partially prepared, this should return true ONLY if fully prepared
12
12
  * TODO: [🧿] Maybe do same process with same granularity and subfinctions as `preparePipeline`
13
13
  * - [🏍] ? Is context in each template
14
- * - [♨] Are samples prepared
14
+ * - [♨] Are examples prepared
15
15
  * - [♨] Are templates prepared
16
16
  */
@@ -24,7 +24,7 @@ export {};
24
24
  /**
25
25
  * TODO: [🧠] Add context to each template (if missing)
26
26
  * TODO: [🧠] What is better name `prepareTemplate` or `prepareTemplateAndParameters`
27
- * TODO: [♨][main] !!! Prepare index the samples and maybe templates
27
+ * TODO: [♨][main] !!! Prepare index the examples and maybe templates
28
28
  * TODO: Write tests for `preparePipeline`
29
29
  * TODO: [🏏] Leverage the batch API and build queues @see https://platform.openai.com/docs/guides/batch
30
30
  * TODO: [🧊] In future one preparation can take data from previous preparation and save tokens and time
@@ -1,4 +1,4 @@
1
1
  export {};
2
2
  /**
3
- * TODO: [📓] Maybe test all file in samples (not just 10-simple.docx)
3
+ * TODO: [📓] Maybe test all file in examples (not just 10-simple.docx)
4
4
  */
@@ -1,4 +1,4 @@
1
1
  export {};
2
2
  /**
3
- * TODO: [📓] Maybe test all file in samples (not just 10-simple.doc)
3
+ * TODO: [📓] Maybe test all file in examples (not just 10-simple.doc)
4
4
  */
@@ -1,4 +1,4 @@
1
1
  export {};
2
2
  /**
3
- * TODO: [📓] Maybe test all file in samples (not just 10-simple.md)
3
+ * TODO: [📓] Maybe test all file in examples (not just 10-simple.md)
4
4
  */
@@ -27,10 +27,10 @@ export type ParameterJson = {
27
27
  */
28
28
  readonly description?: string_markdown_text;
29
29
  /**
30
- * Sample values of the parameter
30
+ * Example values of the parameter
31
31
  * Note: This values won't be actually used as some default values, but they are just for better understanding of the parameter
32
32
  */
33
- readonly sampleValues?: Array<string_parameter_value>;
33
+ readonly exampleValues?: Array<string_parameter_value>;
34
34
  };
35
35
  /**
36
36
  * TODO: [🧠] Should be here registered subparameters from foreach or not?
@@ -18,5 +18,5 @@ export type PreparationJson = {
18
18
  /**
19
19
  * TODO: [🍙] Make some standard order of json properties
20
20
  * TODO: Maybe put here used `modelName`
21
- * TODO: [🍥] When using `date` it changes all samples .ptbk.json files each time so until some more elegant solution omit the time from prepared pipeline
21
+ * TODO: [🍥] When using `date` it changes all examples .ptbk.json files each time so until some more elegant solution omit the time from prepared pipeline
22
22
  */
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@promptbook/node",
3
- "version": "0.72.0",
3
+ "version": "0.73.0",
4
4
  "description": "Supercharge your use of large language models",
5
5
  "private": false,
6
6
  "sideEffects": false,
@@ -31,7 +31,8 @@
31
31
  "o1",
32
32
  "o1-mini",
33
33
  "o1-preview",
34
- "anthropic"
34
+ "anthropic",
35
+ "LLMOps"
35
36
  ],
36
37
  "license": "CC-BY-4.0",
37
38
  "bugs": {
@@ -52,7 +53,7 @@
52
53
  "module": "./esm/index.es.js",
53
54
  "typings": "./esm/typings/src/_packages/node.index.d.ts",
54
55
  "peerDependencies": {
55
- "@promptbook/core": "0.72.0"
56
+ "@promptbook/core": "0.73.0"
56
57
  },
57
58
  "dependencies": {
58
59
  "colors": "1.4.0",
package/umd/index.umd.js CHANGED
@@ -35,7 +35,7 @@
35
35
  /**
36
36
  * The version of the Promptbook library
37
37
  */
38
- var PROMPTBOOK_VERSION = '0.72.0-34';
38
+ var PROMPTBOOK_VERSION = '0.72.0';
39
39
  // TODO: [main] !!!! List here all the versions and annotate + put into script
40
40
 
41
41
  /*! *****************************************************************************
@@ -423,10 +423,10 @@
423
423
  'content',
424
424
  'context',
425
425
  'knowledge',
426
- 'samples',
426
+ 'examples',
427
427
  'modelName',
428
428
  'currentDate',
429
- // <- TODO: !!!!! list here all command names
429
+ // <- TODO: list here all command names
430
430
  // <- TODO: Add more like 'date', 'modelName',...
431
431
  // <- TODO: Add [emoji] + instructions ACRY when adding new reserved parameter
432
432
  ]);
@@ -1360,7 +1360,7 @@
1360
1360
  * TODO: [👷‍♂️] @@@ Manual about construction of llmTools
1361
1361
  */
1362
1362
 
1363
- var PipelineCollection = [{title:"Prepare Knowledge from Markdown",pipelineUrl:"https://promptbook.studio/promptbook/prepare-knowledge-from-markdown.ptbk.md",parameters:[{name:"knowledgeContent",description:"Markdown document content",isInput:true,isOutput:false},{name:"knowledgePieces",description:"The knowledge JSON object",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"knowledge",title:"Knowledge",content:"You are experienced data researcher, extract the important knowledge from the document.\n\n# Rules\n\n- Make pieces of information concise, clear, and easy to understand\n- One piece of information should be approximately 1 paragraph\n- Divide the paragraphs by markdown horizontal lines ---\n- Omit irrelevant information\n- Group redundant information\n- Write just extracted information, nothing else\n\n# The document\n\nTake information from this document:\n\n> {knowledgeContent}",resultingParameterName:"knowledgePieces",dependentParameterNames:["knowledgeContent"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-knowledge-from-markdown.ptbk.md"},{title:"Prepare Keywords",pipelineUrl:"https://promptbook.studio/promptbook/prepare-knowledge-keywords.ptbk.md",parameters:[{name:"knowledgePieceContent",description:"The content",isInput:true,isOutput:false},{name:"keywords",description:"Keywords separated by comma",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"knowledge",title:"Knowledge",content:"You are experienced data researcher, detect the important keywords in the document.\n\n# Rules\n\n- Write just keywords separated by comma\n\n# The document\n\nTake information from this document:\n\n> {knowledgePieceContent}",resultingParameterName:"keywords",dependentParameterNames:["knowledgePieceContent"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-knowledge-keywords.ptbk.md"},{title:"Prepare Title",pipelineUrl:"https://promptbook.studio/promptbook/prepare-knowledge-title.ptbk.md",parameters:[{name:"knowledgePieceContent",description:"The content",isInput:true,isOutput:false},{name:"title",description:"The title of the document",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"knowledge",title:"Knowledge",content:"You are experienced content creator, write best title for the document.\n\n# Rules\n\n- Write just title, nothing else\n- Title should be concise and clear\n- Write maximum 5 words for the title\n\n# The document\n\n> {knowledgePieceContent}",resultingParameterName:"title",expectations:{words:{min:1,max:8}},dependentParameterNames:["knowledgePieceContent"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-knowledge-title.ptbk.md"},{title:"Prepare Keywords",pipelineUrl:"https://promptbook.studio/promptbook/prepare-persona.ptbk.md",parameters:[{name:"availableModelNames",description:"List of available model names separated by comma (,)",isInput:true,isOutput:false},{name:"personaDescription",description:"Description of the persona",isInput:true,isOutput:false},{name:"modelRequirements",description:"Specific requirements for the model",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"make-model-requirements",title:"Make modelRequirements",content:"You are experienced AI engineer, you need to create virtual assistant.\nWrite\n\n## Sample\n\n```json\n{\n\"modelName\": \"gpt-4o\",\n\"systemMessage\": \"You are experienced AI engineer and helpfull assistant.\",\n\"temperature\": 0.7\n}\n```\n\n## Instructions\n\n- Your output format is JSON object\n- Write just the JSON object, no other text should be present\n- It contains the following keys:\n - `modelName`: The name of the model to use\n - `systemMessage`: The system message to provide context to the model\n - `temperature`: The sampling temperature to use\n\n### Key `modelName`\n\nPick from the following models:\n\n- {availableModelNames}\n\n### Key `systemMessage`\n\nThe system message is used to communicate instructions or provide context to the model at the beginning of a conversation. It is displayed in a different format compared to user messages, helping the model understand its role in the conversation. The system message typically guides the model's behavior, sets the tone, or specifies desired output from the model. By utilizing the system message effectively, users can steer the model towards generating more accurate and relevant responses.\n\nFor example:\n\n> You are an experienced AI engineer and helpful assistant.\n\n> You are a friendly and knowledgeable chatbot.\n\n### Key `temperature`\n\nThe sampling temperature, between 0 and 1. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. If set to 0, the model will use log probability to automatically increase the temperature until certain thresholds are hit.\n\nYou can pick a value between 0 and 2. For example:\n\n- `0.1`: Low temperature, extremely conservative and deterministic\n- `0.5`: Medium temperature, balanced between conservative and creative\n- `1.0`: High temperature, creative and bit random\n- `1.5`: Very high temperature, extremely creative and often chaotic and unpredictable\n- `2.0`: Maximum temperature, completely random and unpredictable, for some extreme creative use cases\n\n# The assistant\n\nTake this description of the persona:\n\n> {personaDescription}",resultingParameterName:"modelRequirements",format:"JSON",dependentParameterNames:["availableModelNames","personaDescription"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-persona.ptbk.md"}];
1363
+ var PipelineCollection = [{title:"Prepare Knowledge from Markdown",pipelineUrl:"https://promptbook.studio/promptbook/prepare-knowledge-from-markdown.ptbk.md",parameters:[{name:"knowledgeContent",description:"Markdown document content",isInput:true,isOutput:false},{name:"knowledgePieces",description:"The knowledge JSON object",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"knowledge",title:"Knowledge",content:"You are experienced data researcher, extract the important knowledge from the document.\n\n# Rules\n\n- Make pieces of information concise, clear, and easy to understand\n- One piece of information should be approximately 1 paragraph\n- Divide the paragraphs by markdown horizontal lines ---\n- Omit irrelevant information\n- Group redundant information\n- Write just extracted information, nothing else\n\n# The document\n\nTake information from this document:\n\n> {knowledgeContent}",resultingParameterName:"knowledgePieces",dependentParameterNames:["knowledgeContent"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-knowledge-from-markdown.ptbk.md"},{title:"Prepare Keywords",pipelineUrl:"https://promptbook.studio/promptbook/prepare-knowledge-keywords.ptbk.md",parameters:[{name:"knowledgePieceContent",description:"The content",isInput:true,isOutput:false},{name:"keywords",description:"Keywords separated by comma",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"knowledge",title:"Knowledge",content:"You are experienced data researcher, detect the important keywords in the document.\n\n# Rules\n\n- Write just keywords separated by comma\n\n# The document\n\nTake information from this document:\n\n> {knowledgePieceContent}",resultingParameterName:"keywords",dependentParameterNames:["knowledgePieceContent"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-knowledge-keywords.ptbk.md"},{title:"Prepare Title",pipelineUrl:"https://promptbook.studio/promptbook/prepare-knowledge-title.ptbk.md",parameters:[{name:"knowledgePieceContent",description:"The content",isInput:true,isOutput:false},{name:"title",description:"The title of the document",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"knowledge",title:"Knowledge",content:"You are experienced content creator, write best title for the document.\n\n# Rules\n\n- Write just title, nothing else\n- Title should be concise and clear\n- Write maximum 5 words for the title\n\n# The document\n\n> {knowledgePieceContent}",resultingParameterName:"title",expectations:{words:{min:1,max:8}},dependentParameterNames:["knowledgePieceContent"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-knowledge-title.ptbk.md"},{title:"Prepare Keywords",pipelineUrl:"https://promptbook.studio/promptbook/prepare-persona.ptbk.md",parameters:[{name:"availableModelNames",description:"List of available model names separated by comma (,)",isInput:true,isOutput:false},{name:"personaDescription",description:"Description of the persona",isInput:true,isOutput:false},{name:"modelRequirements",description:"Specific requirements for the model",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"make-model-requirements",title:"Make modelRequirements",content:"You are experienced AI engineer, you need to create virtual assistant.\nWrite\n\n## Example\n\n```json\n{\n\"modelName\": \"gpt-4o\",\n\"systemMessage\": \"You are experienced AI engineer and helpfull assistant.\",\n\"temperature\": 0.7\n}\n```\n\n## Instructions\n\n- Your output format is JSON object\n- Write just the JSON object, no other text should be present\n- It contains the following keys:\n - `modelName`: The name of the model to use\n - `systemMessage`: The system message to provide context to the model\n - `temperature`: The sampling temperature to use\n\n### Key `modelName`\n\nPick from the following models:\n\n- {availableModelNames}\n\n### Key `systemMessage`\n\nThe system message is used to communicate instructions or provide context to the model at the beginning of a conversation. It is displayed in a different format compared to user messages, helping the model understand its role in the conversation. The system message typically guides the model's behavior, sets the tone, or specifies desired output from the model. By utilizing the system message effectively, users can steer the model towards generating more accurate and relevant responses.\n\nFor example:\n\n> You are an experienced AI engineer and helpful assistant.\n\n> You are a friendly and knowledgeable chatbot.\n\n### Key `temperature`\n\nThe sampling temperature, between 0 and 1. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. If set to 0, the model will use log probability to automatically increase the temperature until certain thresholds are hit.\n\nYou can pick a value between 0 and 2. For example:\n\n- `0.1`: Low temperature, extremely conservative and deterministic\n- `0.5`: Medium temperature, balanced between conservative and creative\n- `1.0`: High temperature, creative and bit random\n- `1.5`: Very high temperature, extremely creative and often chaotic and unpredictable\n- `2.0`: Maximum temperature, completely random and unpredictable, for some extreme creative use cases\n\n# The assistant\n\nTake this description of the persona:\n\n> {personaDescription}",resultingParameterName:"modelRequirements",format:"JSON",dependentParameterNames:["availableModelNames","personaDescription"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-persona.ptbk.md"}];
1364
1364
 
1365
1365
  /**
1366
1366
  * This error indicates that the promptbook in a markdown format cannot be parsed into a valid promptbook object
@@ -1787,7 +1787,7 @@
1787
1787
  }
1788
1788
  }
1789
1789
  /**
1790
- * TODO: !!!!! [🧞‍♀️] Do not allow joker + foreach
1790
+ * TODO: !! [🧞‍♀️] Do not allow joker + foreach
1791
1791
  * TODO: [🧠] Work with promptbookVersion
1792
1792
  * TODO: Use here some json-schema, Zod or something similar and change it to:
1793
1793
  * > /**
@@ -1799,7 +1799,7 @@
1799
1799
  * > ex port function validatePipeline(promptbook: really_unknown): asserts promptbook is PipelineJson {
1800
1800
  */
1801
1801
  /**
1802
- * TODO: [🧳][main] !!!! Validate that all samples match expectations
1802
+ * TODO: [🧳][main] !!!! Validate that all examples match expectations
1803
1803
  * TODO: [🧳][🐝][main] !!!! Validate that knowledge is valid (non-void)
1804
1804
  * TODO: [🧳][main] !!!! Validate that persona can be used only with CHAT variant
1805
1805
  * TODO: [🧳][main] !!!! Validate that parameter with reserved name not used RESERVED_PARAMETER_NAMES
@@ -2169,12 +2169,12 @@
2169
2169
  return true;
2170
2170
  }
2171
2171
  /**
2172
- * TODO: [🔃][main] !!!!! If the pipeline was prepared with different version or different set of models, prepare it once again
2172
+ * TODO: [🔃][main] !! If the pipeline was prepared with different version or different set of models, prepare it once again
2173
2173
  * TODO: [🐠] Maybe base this on `makeValidator`
2174
2174
  * TODO: [🧊] Pipeline can be partially prepared, this should return true ONLY if fully prepared
2175
2175
  * TODO: [🧿] Maybe do same process with same granularity and subfinctions as `preparePipeline`
2176
2176
  * - [🏍] ? Is context in each template
2177
- * - [♨] Are samples prepared
2177
+ * - [♨] Are examples prepared
2178
2178
  * - [♨] Are templates prepared
2179
2179
  */
2180
2180
 
@@ -3918,7 +3918,7 @@
3918
3918
  var preparedPipeline, template;
3919
3919
  return __generator(this, function (_a) {
3920
3920
  preparedPipeline = options.preparedPipeline, template = options.template;
3921
- // TODO: [♨] Implement Better - use real index and keyword search from `template` and {samples}
3921
+ // TODO: [♨] Implement Better - use real index and keyword search from `template` and {examples}
3922
3922
  TODO_USE(template);
3923
3923
  return [2 /*return*/, preparedPipeline.knowledgePieces.map(function (_a) {
3924
3924
  var content = _a.content;
@@ -3933,7 +3933,7 @@
3933
3933
  *
3934
3934
  * @private internal utility of `createPipelineExecutor`
3935
3935
  */
3936
- function getSamplesForTemplate(template) {
3936
+ function getExamplesForTemplate(template) {
3937
3937
  return __awaiter(this, void 0, void 0, function () {
3938
3938
  return __generator(this, function (_a) {
3939
3939
  // TODO: [♨] Implement Better - use real index and keyword search
@@ -3950,7 +3950,7 @@
3950
3950
  */
3951
3951
  function getReservedParametersForTemplate(options) {
3952
3952
  return __awaiter(this, void 0, void 0, function () {
3953
- var preparedPipeline, template, pipelineIdentification, context, knowledge, samples, currentDate, modelName, reservedParameters, _loop_1, RESERVED_PARAMETER_NAMES_1, RESERVED_PARAMETER_NAMES_1_1, parameterName;
3953
+ var preparedPipeline, template, pipelineIdentification, context, knowledge, examples, currentDate, modelName, reservedParameters, _loop_1, RESERVED_PARAMETER_NAMES_1, RESERVED_PARAMETER_NAMES_1_1, parameterName;
3954
3954
  var e_1, _a;
3955
3955
  return __generator(this, function (_b) {
3956
3956
  switch (_b.label) {
@@ -3962,16 +3962,16 @@
3962
3962
  return [4 /*yield*/, getKnowledgeForTemplate({ preparedPipeline: preparedPipeline, template: template })];
3963
3963
  case 2:
3964
3964
  knowledge = _b.sent();
3965
- return [4 /*yield*/, getSamplesForTemplate(template)];
3965
+ return [4 /*yield*/, getExamplesForTemplate(template)];
3966
3966
  case 3:
3967
- samples = _b.sent();
3967
+ examples = _b.sent();
3968
3968
  currentDate = new Date().toISOString();
3969
3969
  modelName = RESERVED_PARAMETER_MISSING_VALUE;
3970
3970
  reservedParameters = {
3971
3971
  content: RESERVED_PARAMETER_RESTRICTED,
3972
3972
  context: context,
3973
3973
  knowledge: knowledge,
3974
- samples: samples,
3974
+ examples: examples,
3975
3975
  currentDate: currentDate,
3976
3976
  modelName: modelName,
3977
3977
  };
@@ -4589,7 +4589,7 @@
4589
4589
  });
4590
4590
  }
4591
4591
  /**
4592
- * TODO: [🔃][main] !!!!! If the persona was prepared with different version or different set of models, prepare it once again
4592
+ * TODO: [🔃][main] !! If the persona was prepared with different version or different set of models, prepare it once again
4593
4593
  * TODO: [🏢] !! Check validity of `modelName` in pipeline
4594
4594
  * TODO: [🏢] !! Check validity of `systemMessage` in pipeline
4595
4595
  * TODO: [🏢] !! Check validity of `temperature` in pipeline
@@ -5340,7 +5340,7 @@
5340
5340
  case 0:
5341
5341
  _a = options.maxParallelCount, maxParallelCount = _a === void 0 ? DEFAULT_MAX_PARALLEL_COUNT : _a;
5342
5342
  templates = pipeline.templates, parameters = pipeline.parameters, knowledgePiecesCount = pipeline.knowledgePiecesCount;
5343
- // TODO: [main] !!!!! Apply samples to each template (if missing and is for the template defined)
5343
+ // TODO: [main] !! Apply examples to each template (if missing and is for the template defined)
5344
5344
  TODO_USE(parameters);
5345
5345
  templatesPrepared = new Array(templates.length);
5346
5346
  return [4 /*yield*/, forEachAsync(templates, { maxParallelCount: maxParallelCount /* <- TODO: [🪂] When there are subtasks, this maximul limit can be broken */ }, function (template, index) { return __awaiter(_this, void 0, void 0, function () {
@@ -5370,7 +5370,7 @@
5370
5370
  /**
5371
5371
  * TODO: [🧠] Add context to each template (if missing)
5372
5372
  * TODO: [🧠] What is better name `prepareTemplate` or `prepareTemplateAndParameters`
5373
- * TODO: [♨][main] !!! Prepare index the samples and maybe templates
5373
+ * TODO: [♨][main] !!! Prepare index the examples and maybe templates
5374
5374
  * TODO: Write tests for `preparePipeline`
5375
5375
  * TODO: [🏏] Leverage the batch API and build queues @see https://platform.openai.com/docs/guides/batch
5376
5376
  * TODO: [🧊] In future one preparation can take data from previous preparation and save tokens and time
@@ -5579,7 +5579,7 @@
5579
5579
  'SIMPLE_TEMPLATE',
5580
5580
  'SCRIPT_TEMPLATE',
5581
5581
  'DIALOG_TEMPLATE',
5582
- 'SAMPLE',
5582
+ 'EXAMPLE',
5583
5583
  'KNOWLEDGE',
5584
5584
  'INSTRUMENT',
5585
5585
  'ACTION',
@@ -5638,7 +5638,7 @@
5638
5638
  'SCRIPT',
5639
5639
  'DIALOG',
5640
5640
  // <- [🅱]
5641
- 'SAMPLE',
5641
+ 'EXAMPLE',
5642
5642
  'KNOWLEDGE',
5643
5643
  'INSTRUMENT',
5644
5644
  'ACTION',
@@ -5671,7 +5671,7 @@
5671
5671
  */
5672
5672
  parse: function (input) {
5673
5673
  var normalized = input.normalized;
5674
- normalized = normalized.split('EXAMPLE').join('SAMPLE');
5674
+ normalized = normalized.split('SAMPLE').join('EXAMPLE');
5675
5675
  var templateTypes = TemplateTypes.filter(function (templateType) {
5676
5676
  return normalized.includes(templateType.split('_TEMPLATE').join(''));
5677
5677
  });
@@ -5704,14 +5704,14 @@
5704
5704
  if ($templateJson.content === undefined) {
5705
5705
  throw new UnexpectedError("Content is missing in the templateJson - probbably commands are applied in wrong order");
5706
5706
  }
5707
- if (command.templateType === 'SAMPLE') {
5707
+ if (command.templateType === 'EXAMPLE') {
5708
5708
  expectResultingParameterName();
5709
5709
  var parameter = $pipelineJson.parameters.find(function (param) { return param.name === $templateJson.resultingParameterName; });
5710
5710
  if (parameter === undefined) {
5711
- throw new ParseError("Can not find parameter {".concat($templateJson.resultingParameterName, "} to assign sample value on it"));
5711
+ throw new ParseError("Can not find parameter {".concat($templateJson.resultingParameterName, "} to assign example value on it"));
5712
5712
  }
5713
- parameter.sampleValues = parameter.sampleValues || [];
5714
- parameter.sampleValues.push($templateJson.content);
5713
+ parameter.exampleValues = parameter.exampleValues || [];
5714
+ parameter.exampleValues.push($templateJson.content);
5715
5715
  $templateJson.isTemplate = false;
5716
5716
  return;
5717
5717
  }
@@ -9774,7 +9774,7 @@
9774
9774
  return pipelineJsonStringified;
9775
9775
  }
9776
9776
  /**
9777
- * TODO: [🐝] Not Working propperly @see https://promptbook.studio/samples/mixed-knowledge.ptbk.md
9777
+ * TODO: [🐝] Not Working propperly @see https://promptbook.studio/examples/mixed-knowledge.ptbk.md
9778
9778
  * TODO: [🧠][0] Maybe rename to `stringifyPipelineJson`, `stringifyIndexedJson`,...
9779
9779
  * TODO: [🧠] Maybe more elegant solution than replacing via regex
9780
9780
  * TODO: [🍙] Make some standard order of json properties