@promptbook/core 0.86.5 โ†’ 0.86.8

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@promptbook/core",
3
- "version": "0.86.5",
3
+ "version": "0.86.8",
4
4
  "description": "It's time for a paradigm shift. The future of software in plain English, French or Latin",
5
5
  "private": false,
6
6
  "sideEffects": false,
package/umd/index.umd.js CHANGED
@@ -27,7 +27,7 @@
27
27
  * @generated
28
28
  * @see https://github.com/webgptorg/promptbook
29
29
  */
30
- var PROMPTBOOK_ENGINE_VERSION = '0.86.5';
30
+ var PROMPTBOOK_ENGINE_VERSION = '0.86.8';
31
31
  /**
32
32
  * TODO: string_promptbook_version should be constrained to the all versions of Promptbook engine
33
33
  * Note: [๐Ÿ’ž] Ignore a discrepancy between file name and entity name
@@ -6266,7 +6266,9 @@
6266
6266
  return __generator(this, function (_a) {
6267
6267
  dependentParameterNames = task.dependentParameterNames;
6268
6268
  preparedContent = undefined;
6269
- if (knowledgePiecesCount > 0 && !dependentParameterNames.includes('knowledge')) {
6269
+ if (task.taskType === 'PROMPT_TASK' &&
6270
+ knowledgePiecesCount > 0 &&
6271
+ !dependentParameterNames.includes('knowledge')) {
6270
6272
  preparedContent = spaceTrim.spaceTrim("\n {content}\n\n ## Knowledge\n\n {knowledge}\n ");
6271
6273
  // <- TODO: [๐Ÿง ][๐Ÿงป] Cutomize shape/language/formatting of the addition to the prompt
6272
6274
  dependentParameterNames = __spreadArray(__spreadArray([], __read(dependentParameterNames), false), [
@@ -10814,13 +10816,14 @@
10814
10816
  return /* not await */ llmTools.listModels();
10815
10817
  } });
10816
10818
  var callCommonModel = function (prompt) { return __awaiter(_this, void 0, void 0, function () {
10817
- var key, cacheItem, _a, promptResult, _b;
10819
+ var parameters, content, modelRequirements, key, cacheItem, _a, promptResult, _b;
10818
10820
  return __generator(this, function (_c) {
10819
10821
  switch (_c.label) {
10820
10822
  case 0:
10823
+ parameters = prompt.parameters, content = prompt.content, modelRequirements = prompt.modelRequirements;
10821
10824
  key = titleToName(prompt.title.substring(0, MAX_FILENAME_LENGTH - 10) +
10822
10825
  '-' +
10823
- sha256__default["default"](hexEncoder__default["default"].parse(JSON.stringify(prompt.parameters))).toString( /* hex */));
10826
+ sha256__default["default"](hexEncoder__default["default"].parse(JSON.stringify({ parameters: parameters, content: content, modelRequirements: modelRequirements }))).toString( /* hex */));
10824
10827
  if (!!isCacheReloaded) return [3 /*break*/, 2];
10825
10828
  return [4 /*yield*/, storage.getItem(key)];
10826
10829
  case 1: