@promptbook/core 0.72.0 → 0.73.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -12,4 +12,4 @@ export type TemplateType = TupleToUnion<typeof TemplateTypes>;
12
12
  * @see https://github.com/webgptorg/promptbook#template-type
13
13
  * @public exported from `@promptbook/core`
14
14
  */
15
- export declare const TemplateTypes: readonly ["PROMPT_TEMPLATE", "SIMPLE_TEMPLATE", "SCRIPT_TEMPLATE", "DIALOG_TEMPLATE", "SAMPLE", "KNOWLEDGE", "INSTRUMENT", "ACTION"];
15
+ export declare const TemplateTypes: readonly ["PROMPT_TEMPLATE", "SIMPLE_TEMPLATE", "SCRIPT_TEMPLATE", "DIALOG_TEMPLATE", "EXAMPLE", "KNOWLEDGE", "INSTRUMENT", "ACTION"];
@@ -121,7 +121,7 @@ export declare const REPLACING_NONCE = "u$k42k%!V2zo34w7Fu#@QUHYPW";
121
121
  *
122
122
  * @public exported from `@promptbook/core`
123
123
  */
124
- export declare const RESERVED_PARAMETER_NAMES: readonly ["content", "context", "knowledge", "samples", "modelName", "currentDate"];
124
+ export declare const RESERVED_PARAMETER_NAMES: readonly ["content", "context", "knowledge", "examples", "modelName", "currentDate"];
125
125
  /**
126
126
  * @@@
127
127
  *
@@ -22,9 +22,9 @@ export type renderPipelineMermaidOptions = {
22
22
  */
23
23
  export declare function renderPromptbookMermaid(pipelineJson: PipelineJson, options?: renderPipelineMermaidOptions): string;
24
24
  /**
25
- * TODO: !!!!! FOREACH in mermaid graph
26
- * TODO: !!!!! Knowledge in mermaid graph
27
- * TODO: !!!!! Personas in mermaid graph
25
+ * TODO: [🧠] !! FOREACH in mermaid graph
26
+ * TODO: [🧠] !! Knowledge in mermaid graph
27
+ * TODO: [🧠] !! Personas in mermaid graph
28
28
  * TODO: Maybe use some Mermaid package instead of string templating
29
29
  * TODO: [🕌] When more than 2 functionalities, split into separate functions
30
30
  */
@@ -9,7 +9,7 @@ import type { string_json } from '../../types/typeAliases';
9
9
  */
10
10
  export declare function stringifyPipelineJson<TType>(pipeline: TType): string_json<TType>;
11
11
  /**
12
- * TODO: [🐝] Not Working propperly @see https://promptbook.studio/samples/mixed-knowledge.ptbk.md
12
+ * TODO: [🐝] Not Working propperly @see https://promptbook.studio/examples/mixed-knowledge.ptbk.md
13
13
  * TODO: [🧠][0] Maybe rename to `stringifyPipelineJson`, `stringifyIndexedJson`,...
14
14
  * TODO: [🧠] Maybe more elegant solution than replacing via regex
15
15
  * TODO: [🍙] Make some standard order of json properties
@@ -7,7 +7,7 @@ import type { string_json } from '../../types/typeAliases';
7
7
  * Note: Using here custom import to work in jest tests
8
8
  * Note: Using sync version is 💩 in the production code, but it's ok here in tests
9
9
  *
10
- * @param path - The path to the file relative to samples/pipelines directory
10
+ * @param path - The path to the file relative to examples/pipelines directory
11
11
  * @private internal function of tests
12
12
  */
13
13
  export declare function importPipelineWithoutPreparation(path: `${string}.ptbk.md`): PipelineString;
@@ -20,7 +20,7 @@ export declare function validatePipeline(pipeline: PipelineJson): PipelineJson;
20
20
  */
21
21
  export declare function validatePipelineCore(pipeline: PipelineJson): void;
22
22
  /**
23
- * TODO: !!!!! [🧞‍♀️] Do not allow joker + foreach
23
+ * TODO: !! [🧞‍♀️] Do not allow joker + foreach
24
24
  * TODO: [🧠] Work with promptbookVersion
25
25
  * TODO: Use here some json-schema, Zod or something similar and change it to:
26
26
  * > /**
@@ -32,7 +32,7 @@ export declare function validatePipelineCore(pipeline: PipelineJson): void;
32
32
  * > ex port function validatePipeline(promptbook: really_unknown): asserts promptbook is PipelineJson {
33
33
  */
34
34
  /**
35
- * TODO: [🧳][main] !!!! Validate that all samples match expectations
35
+ * TODO: [🧳][main] !!!! Validate that all examples match expectations
36
36
  * TODO: [🧳][🐝][main] !!!! Validate that knowledge is valid (non-void)
37
37
  * TODO: [🧳][main] !!!! Validate that persona can be used only with CHAT variant
38
38
  * TODO: [🧳][main] !!!! Validate that parameter with reserved name not used RESERVED_PARAMETER_NAMES
@@ -7,4 +7,4 @@ import type { string_parameter_value } from '../../types/typeAliases';
7
7
  *
8
8
  * @private internal utility of `createPipelineExecutor`
9
9
  */
10
- export declare function getSamplesForTemplate(template: ReadonlyDeep<TemplateJson>): Promise<string_parameter_value & string_markdown>;
10
+ export declare function getExamplesForTemplate(template: ReadonlyDeep<TemplateJson>): Promise<string_parameter_value & string_markdown>;
@@ -15,7 +15,7 @@ export type FormatDefinition<TValue extends TPartialValue, TPartialValue extends
15
15
  /**
16
16
  * The name of the format used in .ptbk.md files
17
17
  *
18
- * @sample "JSON"
18
+ * @example "JSON"
19
19
  */
20
20
  readonly formatName: string_name & string_SCREAMING_CASE;
21
21
  /**
@@ -25,7 +25,7 @@ export type FormatDefinition<TValue extends TPartialValue, TPartialValue extends
25
25
  /**
26
26
  * The mime type of the format (if any)
27
27
  *
28
- * @sample "application/json"
28
+ * @example "application/json"
29
29
  */
30
30
  readonly mimeType?: string_mime_type;
31
31
  /**
@@ -11,7 +11,7 @@ export type FormatSubvalueDefinition<TValue extends string, TSettings extends em
11
11
  /**
12
12
  * The name of the format used in .ptbk.md files
13
13
  *
14
- * @sample "CELL"
14
+ * @example "CELL"
15
15
  */
16
16
  readonly subvalueName: string_name & string_SCREAMING_CASE;
17
17
  /**
@@ -10,7 +10,7 @@ import type { string_persona_description } from '../types/typeAliases';
10
10
  */
11
11
  export declare function preparePersona(personaDescription: string_persona_description, tools: Pick<ExecutionTools, 'llm'>, options: PrepareAndScrapeOptions): Promise<PersonaPreparedJson['modelRequirements']>;
12
12
  /**
13
- * TODO: [🔃][main] !!!!! If the persona was prepared with different version or different set of models, prepare it once again
13
+ * TODO: [🔃][main] !! If the persona was prepared with different version or different set of models, prepare it once again
14
14
  * TODO: [🏢] !! Check validity of `modelName` in pipeline
15
15
  * TODO: [🏢] !! Check validity of `systemMessage` in pipeline
16
16
  * TODO: [🏢] !! Check validity of `temperature` in pipeline
@@ -6,11 +6,11 @@ import type { PipelineJson } from '../types/PipelineJson/PipelineJson';
6
6
  */
7
7
  export declare function isPipelinePrepared(pipeline: PipelineJson): boolean;
8
8
  /**
9
- * TODO: [🔃][main] !!!!! If the pipeline was prepared with different version or different set of models, prepare it once again
9
+ * TODO: [🔃][main] !! If the pipeline was prepared with different version or different set of models, prepare it once again
10
10
  * TODO: [🐠] Maybe base this on `makeValidator`
11
11
  * TODO: [🧊] Pipeline can be partially prepared, this should return true ONLY if fully prepared
12
12
  * TODO: [🧿] Maybe do same process with same granularity and subfinctions as `preparePipeline`
13
13
  * - [🏍] ? Is context in each template
14
- * - [♨] Are samples prepared
14
+ * - [♨] Are examples prepared
15
15
  * - [♨] Are templates prepared
16
16
  */
@@ -24,7 +24,7 @@ export {};
24
24
  /**
25
25
  * TODO: [🧠] Add context to each template (if missing)
26
26
  * TODO: [🧠] What is better name `prepareTemplate` or `prepareTemplateAndParameters`
27
- * TODO: [♨][main] !!! Prepare index the samples and maybe templates
27
+ * TODO: [♨][main] !!! Prepare index the examples and maybe templates
28
28
  * TODO: Write tests for `preparePipeline`
29
29
  * TODO: [🏏] Leverage the batch API and build queues @see https://platform.openai.com/docs/guides/batch
30
30
  * TODO: [🧊] In future one preparation can take data from previous preparation and save tokens and time
@@ -1,4 +1,4 @@
1
1
  export {};
2
2
  /**
3
- * TODO: [📓] Maybe test all file in samples (not just 10-simple.docx)
3
+ * TODO: [📓] Maybe test all file in examples (not just 10-simple.docx)
4
4
  */
@@ -1,4 +1,4 @@
1
1
  export {};
2
2
  /**
3
- * TODO: [📓] Maybe test all file in samples (not just 10-simple.doc)
3
+ * TODO: [📓] Maybe test all file in examples (not just 10-simple.doc)
4
4
  */
@@ -1,4 +1,4 @@
1
1
  export {};
2
2
  /**
3
- * TODO: [📓] Maybe test all file in samples (not just 10-simple.md)
3
+ * TODO: [📓] Maybe test all file in examples (not just 10-simple.md)
4
4
  */
@@ -27,10 +27,10 @@ export type ParameterJson = {
27
27
  */
28
28
  readonly description?: string_markdown_text;
29
29
  /**
30
- * Sample values of the parameter
30
+ * Example values of the parameter
31
31
  * Note: This values won't be actually used as some default values, but they are just for better understanding of the parameter
32
32
  */
33
- readonly sampleValues?: Array<string_parameter_value>;
33
+ readonly exampleValues?: Array<string_parameter_value>;
34
34
  };
35
35
  /**
36
36
  * TODO: [🧠] Should be here registered subparameters from foreach or not?
@@ -18,5 +18,5 @@ export type PreparationJson = {
18
18
  /**
19
19
  * TODO: [🍙] Make some standard order of json properties
20
20
  * TODO: Maybe put here used `modelName`
21
- * TODO: [🍥] When using `date` it changes all samples .ptbk.json files each time so until some more elegant solution omit the time from prepared pipeline
21
+ * TODO: [🍥] When using `date` it changes all examples .ptbk.json files each time so until some more elegant solution omit the time from prepared pipeline
22
22
  */
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@promptbook/core",
3
- "version": "0.72.0",
3
+ "version": "0.73.0",
4
4
  "description": "Supercharge your use of large language models",
5
5
  "private": false,
6
6
  "sideEffects": false,
@@ -31,7 +31,8 @@
31
31
  "o1",
32
32
  "o1-mini",
33
33
  "o1-preview",
34
- "anthropic"
34
+ "anthropic",
35
+ "LLMOps"
35
36
  ],
36
37
  "license": "CC-BY-4.0",
37
38
  "bugs": {
package/umd/index.umd.js CHANGED
@@ -16,7 +16,7 @@
16
16
  /**
17
17
  * The version of the Promptbook library
18
18
  */
19
- var PROMPTBOOK_VERSION = '0.72.0-34';
19
+ var PROMPTBOOK_VERSION = '0.72.0';
20
20
  // TODO: [main] !!!! List here all the versions and annotate + put into script
21
21
 
22
22
  /*! *****************************************************************************
@@ -727,10 +727,10 @@
727
727
  'content',
728
728
  'context',
729
729
  'knowledge',
730
- 'samples',
730
+ 'examples',
731
731
  'modelName',
732
732
  'currentDate',
733
- // <- TODO: !!!!! list here all command names
733
+ // <- TODO: list here all command names
734
734
  // <- TODO: Add more like 'date', 'modelName',...
735
735
  // <- TODO: Add [emoji] + instructions ACRY when adding new reserved parameter
736
736
  ]);
@@ -1229,7 +1229,7 @@
1229
1229
  }
1230
1230
  }
1231
1231
  /**
1232
- * TODO: !!!!! [🧞‍♀️] Do not allow joker + foreach
1232
+ * TODO: !! [🧞‍♀️] Do not allow joker + foreach
1233
1233
  * TODO: [🧠] Work with promptbookVersion
1234
1234
  * TODO: Use here some json-schema, Zod or something similar and change it to:
1235
1235
  * > /**
@@ -1241,7 +1241,7 @@
1241
1241
  * > ex port function validatePipeline(promptbook: really_unknown): asserts promptbook is PipelineJson {
1242
1242
  */
1243
1243
  /**
1244
- * TODO: [🧳][main] !!!! Validate that all samples match expectations
1244
+ * TODO: [🧳][main] !!!! Validate that all examples match expectations
1245
1245
  * TODO: [🧳][🐝][main] !!!! Validate that knowledge is valid (non-void)
1246
1246
  * TODO: [🧳][main] !!!! Validate that persona can be used only with CHAT variant
1247
1247
  * TODO: [🧳][main] !!!! Validate that parameter with reserved name not used RESERVED_PARAMETER_NAMES
@@ -1653,7 +1653,7 @@
1653
1653
  'SIMPLE_TEMPLATE',
1654
1654
  'SCRIPT_TEMPLATE',
1655
1655
  'DIALOG_TEMPLATE',
1656
- 'SAMPLE',
1656
+ 'EXAMPLE',
1657
1657
  'KNOWLEDGE',
1658
1658
  'INSTRUMENT',
1659
1659
  'ACTION',
@@ -2295,7 +2295,7 @@
2295
2295
  * TODO: [👷‍♂️] @@@ Manual about construction of llmTools
2296
2296
  */
2297
2297
 
2298
- var PipelineCollection = [{title:"Prepare Knowledge from Markdown",pipelineUrl:"https://promptbook.studio/promptbook/prepare-knowledge-from-markdown.ptbk.md",parameters:[{name:"knowledgeContent",description:"Markdown document content",isInput:true,isOutput:false},{name:"knowledgePieces",description:"The knowledge JSON object",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"knowledge",title:"Knowledge",content:"You are experienced data researcher, extract the important knowledge from the document.\n\n# Rules\n\n- Make pieces of information concise, clear, and easy to understand\n- One piece of information should be approximately 1 paragraph\n- Divide the paragraphs by markdown horizontal lines ---\n- Omit irrelevant information\n- Group redundant information\n- Write just extracted information, nothing else\n\n# The document\n\nTake information from this document:\n\n> {knowledgeContent}",resultingParameterName:"knowledgePieces",dependentParameterNames:["knowledgeContent"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-knowledge-from-markdown.ptbk.md"},{title:"Prepare Keywords",pipelineUrl:"https://promptbook.studio/promptbook/prepare-knowledge-keywords.ptbk.md",parameters:[{name:"knowledgePieceContent",description:"The content",isInput:true,isOutput:false},{name:"keywords",description:"Keywords separated by comma",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"knowledge",title:"Knowledge",content:"You are experienced data researcher, detect the important keywords in the document.\n\n# Rules\n\n- Write just keywords separated by comma\n\n# The document\n\nTake information from this document:\n\n> {knowledgePieceContent}",resultingParameterName:"keywords",dependentParameterNames:["knowledgePieceContent"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-knowledge-keywords.ptbk.md"},{title:"Prepare Title",pipelineUrl:"https://promptbook.studio/promptbook/prepare-knowledge-title.ptbk.md",parameters:[{name:"knowledgePieceContent",description:"The content",isInput:true,isOutput:false},{name:"title",description:"The title of the document",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"knowledge",title:"Knowledge",content:"You are experienced content creator, write best title for the document.\n\n# Rules\n\n- Write just title, nothing else\n- Title should be concise and clear\n- Write maximum 5 words for the title\n\n# The document\n\n> {knowledgePieceContent}",resultingParameterName:"title",expectations:{words:{min:1,max:8}},dependentParameterNames:["knowledgePieceContent"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-knowledge-title.ptbk.md"},{title:"Prepare Keywords",pipelineUrl:"https://promptbook.studio/promptbook/prepare-persona.ptbk.md",parameters:[{name:"availableModelNames",description:"List of available model names separated by comma (,)",isInput:true,isOutput:false},{name:"personaDescription",description:"Description of the persona",isInput:true,isOutput:false},{name:"modelRequirements",description:"Specific requirements for the model",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"make-model-requirements",title:"Make modelRequirements",content:"You are experienced AI engineer, you need to create virtual assistant.\nWrite\n\n## Sample\n\n```json\n{\n\"modelName\": \"gpt-4o\",\n\"systemMessage\": \"You are experienced AI engineer and helpfull assistant.\",\n\"temperature\": 0.7\n}\n```\n\n## Instructions\n\n- Your output format is JSON object\n- Write just the JSON object, no other text should be present\n- It contains the following keys:\n - `modelName`: The name of the model to use\n - `systemMessage`: The system message to provide context to the model\n - `temperature`: The sampling temperature to use\n\n### Key `modelName`\n\nPick from the following models:\n\n- {availableModelNames}\n\n### Key `systemMessage`\n\nThe system message is used to communicate instructions or provide context to the model at the beginning of a conversation. It is displayed in a different format compared to user messages, helping the model understand its role in the conversation. The system message typically guides the model's behavior, sets the tone, or specifies desired output from the model. By utilizing the system message effectively, users can steer the model towards generating more accurate and relevant responses.\n\nFor example:\n\n> You are an experienced AI engineer and helpful assistant.\n\n> You are a friendly and knowledgeable chatbot.\n\n### Key `temperature`\n\nThe sampling temperature, between 0 and 1. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. If set to 0, the model will use log probability to automatically increase the temperature until certain thresholds are hit.\n\nYou can pick a value between 0 and 2. For example:\n\n- `0.1`: Low temperature, extremely conservative and deterministic\n- `0.5`: Medium temperature, balanced between conservative and creative\n- `1.0`: High temperature, creative and bit random\n- `1.5`: Very high temperature, extremely creative and often chaotic and unpredictable\n- `2.0`: Maximum temperature, completely random and unpredictable, for some extreme creative use cases\n\n# The assistant\n\nTake this description of the persona:\n\n> {personaDescription}",resultingParameterName:"modelRequirements",format:"JSON",dependentParameterNames:["availableModelNames","personaDescription"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-persona.ptbk.md"}];
2298
+ var PipelineCollection = [{title:"Prepare Knowledge from Markdown",pipelineUrl:"https://promptbook.studio/promptbook/prepare-knowledge-from-markdown.ptbk.md",parameters:[{name:"knowledgeContent",description:"Markdown document content",isInput:true,isOutput:false},{name:"knowledgePieces",description:"The knowledge JSON object",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"knowledge",title:"Knowledge",content:"You are experienced data researcher, extract the important knowledge from the document.\n\n# Rules\n\n- Make pieces of information concise, clear, and easy to understand\n- One piece of information should be approximately 1 paragraph\n- Divide the paragraphs by markdown horizontal lines ---\n- Omit irrelevant information\n- Group redundant information\n- Write just extracted information, nothing else\n\n# The document\n\nTake information from this document:\n\n> {knowledgeContent}",resultingParameterName:"knowledgePieces",dependentParameterNames:["knowledgeContent"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-knowledge-from-markdown.ptbk.md"},{title:"Prepare Keywords",pipelineUrl:"https://promptbook.studio/promptbook/prepare-knowledge-keywords.ptbk.md",parameters:[{name:"knowledgePieceContent",description:"The content",isInput:true,isOutput:false},{name:"keywords",description:"Keywords separated by comma",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"knowledge",title:"Knowledge",content:"You are experienced data researcher, detect the important keywords in the document.\n\n# Rules\n\n- Write just keywords separated by comma\n\n# The document\n\nTake information from this document:\n\n> {knowledgePieceContent}",resultingParameterName:"keywords",dependentParameterNames:["knowledgePieceContent"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-knowledge-keywords.ptbk.md"},{title:"Prepare Title",pipelineUrl:"https://promptbook.studio/promptbook/prepare-knowledge-title.ptbk.md",parameters:[{name:"knowledgePieceContent",description:"The content",isInput:true,isOutput:false},{name:"title",description:"The title of the document",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"knowledge",title:"Knowledge",content:"You are experienced content creator, write best title for the document.\n\n# Rules\n\n- Write just title, nothing else\n- Title should be concise and clear\n- Write maximum 5 words for the title\n\n# The document\n\n> {knowledgePieceContent}",resultingParameterName:"title",expectations:{words:{min:1,max:8}},dependentParameterNames:["knowledgePieceContent"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-knowledge-title.ptbk.md"},{title:"Prepare Keywords",pipelineUrl:"https://promptbook.studio/promptbook/prepare-persona.ptbk.md",parameters:[{name:"availableModelNames",description:"List of available model names separated by comma (,)",isInput:true,isOutput:false},{name:"personaDescription",description:"Description of the persona",isInput:true,isOutput:false},{name:"modelRequirements",description:"Specific requirements for the model",isInput:false,isOutput:true}],templates:[{templateType:"PROMPT_TEMPLATE",name:"make-model-requirements",title:"Make modelRequirements",content:"You are experienced AI engineer, you need to create virtual assistant.\nWrite\n\n## Example\n\n```json\n{\n\"modelName\": \"gpt-4o\",\n\"systemMessage\": \"You are experienced AI engineer and helpfull assistant.\",\n\"temperature\": 0.7\n}\n```\n\n## Instructions\n\n- Your output format is JSON object\n- Write just the JSON object, no other text should be present\n- It contains the following keys:\n - `modelName`: The name of the model to use\n - `systemMessage`: The system message to provide context to the model\n - `temperature`: The sampling temperature to use\n\n### Key `modelName`\n\nPick from the following models:\n\n- {availableModelNames}\n\n### Key `systemMessage`\n\nThe system message is used to communicate instructions or provide context to the model at the beginning of a conversation. It is displayed in a different format compared to user messages, helping the model understand its role in the conversation. The system message typically guides the model's behavior, sets the tone, or specifies desired output from the model. By utilizing the system message effectively, users can steer the model towards generating more accurate and relevant responses.\n\nFor example:\n\n> You are an experienced AI engineer and helpful assistant.\n\n> You are a friendly and knowledgeable chatbot.\n\n### Key `temperature`\n\nThe sampling temperature, between 0 and 1. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. If set to 0, the model will use log probability to automatically increase the temperature until certain thresholds are hit.\n\nYou can pick a value between 0 and 2. For example:\n\n- `0.1`: Low temperature, extremely conservative and deterministic\n- `0.5`: Medium temperature, balanced between conservative and creative\n- `1.0`: High temperature, creative and bit random\n- `1.5`: Very high temperature, extremely creative and often chaotic and unpredictable\n- `2.0`: Maximum temperature, completely random and unpredictable, for some extreme creative use cases\n\n# The assistant\n\nTake this description of the persona:\n\n> {personaDescription}",resultingParameterName:"modelRequirements",format:"JSON",dependentParameterNames:["availableModelNames","personaDescription"]}],knowledgeSources:[],knowledgePieces:[],personas:[],preparations:[],sourceFile:"./promptbook-collection/prepare-persona.ptbk.md"}];
2299
2299
 
2300
2300
  /**
2301
2301
  * This error indicates that the pipeline collection cannot be propperly loaded
@@ -2468,12 +2468,12 @@
2468
2468
  return true;
2469
2469
  }
2470
2470
  /**
2471
- * TODO: [🔃][main] !!!!! If the pipeline was prepared with different version or different set of models, prepare it once again
2471
+ * TODO: [🔃][main] !! If the pipeline was prepared with different version or different set of models, prepare it once again
2472
2472
  * TODO: [🐠] Maybe base this on `makeValidator`
2473
2473
  * TODO: [🧊] Pipeline can be partially prepared, this should return true ONLY if fully prepared
2474
2474
  * TODO: [🧿] Maybe do same process with same granularity and subfinctions as `preparePipeline`
2475
2475
  * - [🏍] ? Is context in each template
2476
- * - [♨] Are samples prepared
2476
+ * - [♨] Are examples prepared
2477
2477
  * - [♨] Are templates prepared
2478
2478
  */
2479
2479
 
@@ -4239,7 +4239,7 @@
4239
4239
  var preparedPipeline, template;
4240
4240
  return __generator(this, function (_a) {
4241
4241
  preparedPipeline = options.preparedPipeline, template = options.template;
4242
- // TODO: [♨] Implement Better - use real index and keyword search from `template` and {samples}
4242
+ // TODO: [♨] Implement Better - use real index and keyword search from `template` and {examples}
4243
4243
  TODO_USE(template);
4244
4244
  return [2 /*return*/, preparedPipeline.knowledgePieces.map(function (_a) {
4245
4245
  var content = _a.content;
@@ -4254,7 +4254,7 @@
4254
4254
  *
4255
4255
  * @private internal utility of `createPipelineExecutor`
4256
4256
  */
4257
- function getSamplesForTemplate(template) {
4257
+ function getExamplesForTemplate(template) {
4258
4258
  return __awaiter(this, void 0, void 0, function () {
4259
4259
  return __generator(this, function (_a) {
4260
4260
  // TODO: [♨] Implement Better - use real index and keyword search
@@ -4271,7 +4271,7 @@
4271
4271
  */
4272
4272
  function getReservedParametersForTemplate(options) {
4273
4273
  return __awaiter(this, void 0, void 0, function () {
4274
- var preparedPipeline, template, pipelineIdentification, context, knowledge, samples, currentDate, modelName, reservedParameters, _loop_1, RESERVED_PARAMETER_NAMES_1, RESERVED_PARAMETER_NAMES_1_1, parameterName;
4274
+ var preparedPipeline, template, pipelineIdentification, context, knowledge, examples, currentDate, modelName, reservedParameters, _loop_1, RESERVED_PARAMETER_NAMES_1, RESERVED_PARAMETER_NAMES_1_1, parameterName;
4275
4275
  var e_1, _a;
4276
4276
  return __generator(this, function (_b) {
4277
4277
  switch (_b.label) {
@@ -4283,16 +4283,16 @@
4283
4283
  return [4 /*yield*/, getKnowledgeForTemplate({ preparedPipeline: preparedPipeline, template: template })];
4284
4284
  case 2:
4285
4285
  knowledge = _b.sent();
4286
- return [4 /*yield*/, getSamplesForTemplate(template)];
4286
+ return [4 /*yield*/, getExamplesForTemplate(template)];
4287
4287
  case 3:
4288
- samples = _b.sent();
4288
+ examples = _b.sent();
4289
4289
  currentDate = new Date().toISOString();
4290
4290
  modelName = RESERVED_PARAMETER_MISSING_VALUE;
4291
4291
  reservedParameters = {
4292
4292
  content: RESERVED_PARAMETER_RESTRICTED,
4293
4293
  context: context,
4294
4294
  knowledge: knowledge,
4295
- samples: samples,
4295
+ examples: examples,
4296
4296
  currentDate: currentDate,
4297
4297
  modelName: modelName,
4298
4298
  };
@@ -4910,7 +4910,7 @@
4910
4910
  });
4911
4911
  }
4912
4912
  /**
4913
- * TODO: [🔃][main] !!!!! If the persona was prepared with different version or different set of models, prepare it once again
4913
+ * TODO: [🔃][main] !! If the persona was prepared with different version or different set of models, prepare it once again
4914
4914
  * TODO: [🏢] !! Check validity of `modelName` in pipeline
4915
4915
  * TODO: [🏢] !! Check validity of `systemMessage` in pipeline
4916
4916
  * TODO: [🏢] !! Check validity of `temperature` in pipeline
@@ -5661,7 +5661,7 @@
5661
5661
  case 0:
5662
5662
  _a = options.maxParallelCount, maxParallelCount = _a === void 0 ? DEFAULT_MAX_PARALLEL_COUNT : _a;
5663
5663
  templates = pipeline.templates, parameters = pipeline.parameters, knowledgePiecesCount = pipeline.knowledgePiecesCount;
5664
- // TODO: [main] !!!!! Apply samples to each template (if missing and is for the template defined)
5664
+ // TODO: [main] !! Apply examples to each template (if missing and is for the template defined)
5665
5665
  TODO_USE(parameters);
5666
5666
  templatesPrepared = new Array(templates.length);
5667
5667
  return [4 /*yield*/, forEachAsync(templates, { maxParallelCount: maxParallelCount /* <- TODO: [🪂] When there are subtasks, this maximul limit can be broken */ }, function (template, index) { return __awaiter(_this, void 0, void 0, function () {
@@ -5691,7 +5691,7 @@
5691
5691
  /**
5692
5692
  * TODO: [🧠] Add context to each template (if missing)
5693
5693
  * TODO: [🧠] What is better name `prepareTemplate` or `prepareTemplateAndParameters`
5694
- * TODO: [♨][main] !!! Prepare index the samples and maybe templates
5694
+ * TODO: [♨][main] !!! Prepare index the examples and maybe templates
5695
5695
  * TODO: Write tests for `preparePipeline`
5696
5696
  * TODO: [🏏] Leverage the batch API and build queues @see https://platform.openai.com/docs/guides/batch
5697
5697
  * TODO: [🧊] In future one preparation can take data from previous preparation and save tokens and time
@@ -5941,7 +5941,7 @@
5941
5941
  'SCRIPT',
5942
5942
  'DIALOG',
5943
5943
  // <- [🅱]
5944
- 'SAMPLE',
5944
+ 'EXAMPLE',
5945
5945
  'KNOWLEDGE',
5946
5946
  'INSTRUMENT',
5947
5947
  'ACTION',
@@ -5974,7 +5974,7 @@
5974
5974
  */
5975
5975
  parse: function (input) {
5976
5976
  var normalized = input.normalized;
5977
- normalized = normalized.split('EXAMPLE').join('SAMPLE');
5977
+ normalized = normalized.split('SAMPLE').join('EXAMPLE');
5978
5978
  var templateTypes = TemplateTypes.filter(function (templateType) {
5979
5979
  return normalized.includes(templateType.split('_TEMPLATE').join(''));
5980
5980
  });
@@ -6007,14 +6007,14 @@
6007
6007
  if ($templateJson.content === undefined) {
6008
6008
  throw new UnexpectedError("Content is missing in the templateJson - probbably commands are applied in wrong order");
6009
6009
  }
6010
- if (command.templateType === 'SAMPLE') {
6010
+ if (command.templateType === 'EXAMPLE') {
6011
6011
  expectResultingParameterName();
6012
6012
  var parameter = $pipelineJson.parameters.find(function (param) { return param.name === $templateJson.resultingParameterName; });
6013
6013
  if (parameter === undefined) {
6014
- throw new ParseError("Can not find parameter {".concat($templateJson.resultingParameterName, "} to assign sample value on it"));
6014
+ throw new ParseError("Can not find parameter {".concat($templateJson.resultingParameterName, "} to assign example value on it"));
6015
6015
  }
6016
- parameter.sampleValues = parameter.sampleValues || [];
6017
- parameter.sampleValues.push($templateJson.content);
6016
+ parameter.exampleValues = parameter.exampleValues || [];
6017
+ parameter.exampleValues.push($templateJson.content);
6018
6018
  $templateJson.isTemplate = false;
6019
6019
  return;
6020
6020
  }
@@ -8602,9 +8602,9 @@
8602
8602
  return promptbookMermaid;
8603
8603
  }
8604
8604
  /**
8605
- * TODO: !!!!! FOREACH in mermaid graph
8606
- * TODO: !!!!! Knowledge in mermaid graph
8607
- * TODO: !!!!! Personas in mermaid graph
8605
+ * TODO: [🧠] !! FOREACH in mermaid graph
8606
+ * TODO: [🧠] !! Knowledge in mermaid graph
8607
+ * TODO: [🧠] !! Personas in mermaid graph
8608
8608
  * TODO: Maybe use some Mermaid package instead of string templating
8609
8609
  * TODO: [🕌] When more than 2 functionalities, split into separate functions
8610
8610
  */
@@ -8704,7 +8704,7 @@
8704
8704
  return pipelineJsonStringified;
8705
8705
  }
8706
8706
  /**
8707
- * TODO: [🐝] Not Working propperly @see https://promptbook.studio/samples/mixed-knowledge.ptbk.md
8707
+ * TODO: [🐝] Not Working propperly @see https://promptbook.studio/examples/mixed-knowledge.ptbk.md
8708
8708
  * TODO: [🧠][0] Maybe rename to `stringifyPipelineJson`, `stringifyIndexedJson`,...
8709
8709
  * TODO: [🧠] Maybe more elegant solution than replacing via regex
8710
8710
  * TODO: [🍙] Make some standard order of json properties