@promptbook/core 0.94.0-7 → 0.94.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +6 -11
- package/esm/index.es.js +11 -7
- package/esm/index.es.js.map +1 -1
- package/esm/typings/src/version.d.ts +1 -1
- package/package.json +24 -1
- package/umd/index.umd.js +11 -7
- package/umd/index.umd.js.map +1 -1
|
@@ -15,7 +15,7 @@ export declare const BOOK_LANGUAGE_VERSION: string_semantic_version;
|
|
|
15
15
|
export declare const PROMPTBOOK_ENGINE_VERSION: string_promptbook_version;
|
|
16
16
|
/**
|
|
17
17
|
* Represents the version string of the Promptbook engine.
|
|
18
|
-
* It follows semantic versioning (e.g., `0.94.0-
|
|
18
|
+
* It follows semantic versioning (e.g., `0.94.0-16`).
|
|
19
19
|
*
|
|
20
20
|
* @generated
|
|
21
21
|
*/
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@promptbook/core",
|
|
3
|
-
"version": "0.94.0
|
|
3
|
+
"version": "0.94.0",
|
|
4
4
|
"description": "Promptbook: Run AI apps in plain human language across multiple models and platforms",
|
|
5
5
|
"private": false,
|
|
6
6
|
"sideEffects": false,
|
|
@@ -36,6 +36,29 @@
|
|
|
36
36
|
"o1-mini",
|
|
37
37
|
"o1-preview",
|
|
38
38
|
"anthropic",
|
|
39
|
+
"claude",
|
|
40
|
+
"claude-3",
|
|
41
|
+
"claude-3-opus",
|
|
42
|
+
"claude-3-sonnet",
|
|
43
|
+
"claude-3-haiku",
|
|
44
|
+
"gemini",
|
|
45
|
+
"gemini-pro",
|
|
46
|
+
"gemini-flash",
|
|
47
|
+
"mixtral",
|
|
48
|
+
"mistral",
|
|
49
|
+
"ollama",
|
|
50
|
+
"ai-orchestration",
|
|
51
|
+
"prompt-engineering",
|
|
52
|
+
"llmops",
|
|
53
|
+
"multimodal",
|
|
54
|
+
"reasoning",
|
|
55
|
+
"rag",
|
|
56
|
+
"embeddings",
|
|
57
|
+
"function-calling",
|
|
58
|
+
"large-language-models",
|
|
59
|
+
"ai-application-framework",
|
|
60
|
+
"text-generation",
|
|
61
|
+
"ai-agents",
|
|
39
62
|
"LLMOps"
|
|
40
63
|
],
|
|
41
64
|
"license": "BUSL-1.1",
|
package/umd/index.umd.js
CHANGED
|
@@ -27,7 +27,7 @@
|
|
|
27
27
|
* @generated
|
|
28
28
|
* @see https://github.com/webgptorg/promptbook
|
|
29
29
|
*/
|
|
30
|
-
const PROMPTBOOK_ENGINE_VERSION = '0.94.0
|
|
30
|
+
const PROMPTBOOK_ENGINE_VERSION = '0.94.0';
|
|
31
31
|
/**
|
|
32
32
|
* TODO: string_promptbook_version should be constrained to the all versions of Promptbook engine
|
|
33
33
|
* Note: [💞] Ignore a discrepancy between file name and entity name
|
|
@@ -10960,7 +10960,7 @@
|
|
|
10960
10960
|
packageName: '@promptbook/azure-openai',
|
|
10961
10961
|
className: 'AzureOpenAiExecutionTools',
|
|
10962
10962
|
options: {
|
|
10963
|
-
apiKey: '
|
|
10963
|
+
apiKey: '',
|
|
10964
10964
|
resourceName: 'my-resource-name',
|
|
10965
10965
|
deploymentName: 'my-deployment-name',
|
|
10966
10966
|
maxRequestsPerMinute: DEFAULT_MAX_REQUESTS_PER_MINUTE,
|
|
@@ -11131,6 +11131,13 @@
|
|
|
11131
11131
|
* Note: [💞] Ignore a discrepancy between file name and entity name
|
|
11132
11132
|
*/
|
|
11133
11133
|
|
|
11134
|
+
/**
|
|
11135
|
+
* Default base URL for Ollama API
|
|
11136
|
+
*
|
|
11137
|
+
* @public exported from `@promptbook/ollama`
|
|
11138
|
+
*/
|
|
11139
|
+
const DEFAULT_OLLAMA_BASE_URL = 'http://localhost:11434/v1';
|
|
11140
|
+
|
|
11134
11141
|
/**
|
|
11135
11142
|
* Registration of LLM provider metadata
|
|
11136
11143
|
*
|
|
@@ -11153,8 +11160,7 @@
|
|
|
11153
11160
|
packageName: '@promptbook/ollama',
|
|
11154
11161
|
className: 'OllamaExecutionTools',
|
|
11155
11162
|
options: {
|
|
11156
|
-
baseURL:
|
|
11157
|
-
model: 'llama2',
|
|
11163
|
+
baseURL: DEFAULT_OLLAMA_BASE_URL,
|
|
11158
11164
|
maxRequestsPerMinute: DEFAULT_MAX_REQUESTS_PER_MINUTE,
|
|
11159
11165
|
},
|
|
11160
11166
|
};
|
|
@@ -11167,8 +11173,6 @@
|
|
|
11167
11173
|
className: 'OllamaExecutionTools',
|
|
11168
11174
|
options: {
|
|
11169
11175
|
baseURL: env.OLLAMA_BASE_URL,
|
|
11170
|
-
model: env.OLLAMA_MODEL || 'llama2',
|
|
11171
|
-
maxRequestsPerMinute: DEFAULT_MAX_REQUESTS_PER_MINUTE,
|
|
11172
11176
|
},
|
|
11173
11177
|
};
|
|
11174
11178
|
}
|
|
@@ -11261,7 +11265,7 @@
|
|
|
11261
11265
|
options: {
|
|
11262
11266
|
apiKey: env.OPENAI_API_KEY!,
|
|
11263
11267
|
assistantId: env.OPENAI_XXX!
|
|
11264
|
-
},
|
|
11268
|
+
} satisfies OpenAiAssistantExecutionToolsOptions,
|
|
11265
11269
|
};
|
|
11266
11270
|
}
|
|
11267
11271
|
|