@promptbook/ollama 0.94.0-7 → 0.95.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +10 -21
- package/esm/index.es.js +192 -64
- package/esm/index.es.js.map +1 -1
- package/esm/typings/src/_packages/types.index.d.ts +2 -2
- package/esm/typings/src/_packages/{wizzard.index.d.ts → wizard.index.d.ts} +2 -2
- package/esm/typings/src/cli/cli-commands/prettify.d.ts +1 -1
- package/esm/typings/src/cli/cli-commands/test-command.d.ts +1 -1
- package/esm/typings/src/conversion/archive/loadArchive.d.ts +1 -1
- package/esm/typings/src/conversion/archive/saveArchive.d.ts +2 -2
- package/esm/typings/src/conversion/prettify/renderPipelineMermaidOptions.d.ts +1 -1
- package/esm/typings/src/dialogs/callback/CallbackInterfaceTools.d.ts +1 -1
- package/esm/typings/src/execution/AbstractTaskResult.d.ts +2 -2
- package/esm/typings/src/execution/createPipelineExecutor/00-CreatePipelineExecutorOptions.d.ts +1 -1
- package/esm/typings/src/execution/execution-report/ExecutionPromptReportJson.d.ts +2 -2
- package/esm/typings/src/execution/translation/automatic-translate/translateMessages.d.ts +1 -1
- package/esm/typings/src/llm-providers/_common/register/{$provideLlmToolsForWizzardOrCli.d.ts → $provideLlmToolsForWizardOrCli.d.ts} +2 -2
- package/esm/typings/src/llm-providers/anthropic-claude/register-configuration.d.ts +1 -1
- package/esm/typings/src/llm-providers/anthropic-claude/register-constructor.d.ts +1 -1
- package/esm/typings/src/llm-providers/azure-openai/register-configuration.d.ts +1 -1
- package/esm/typings/src/llm-providers/azure-openai/register-constructor.d.ts +1 -1
- package/esm/typings/src/llm-providers/deepseek/register-configuration.d.ts +1 -1
- package/esm/typings/src/llm-providers/deepseek/register-constructor.d.ts +1 -1
- package/esm/typings/src/llm-providers/google/register-configuration.d.ts +1 -1
- package/esm/typings/src/llm-providers/google/register-constructor.d.ts +1 -1
- package/esm/typings/src/llm-providers/ollama/register-configuration.d.ts +1 -1
- package/esm/typings/src/llm-providers/ollama/register-constructor.d.ts +1 -1
- package/esm/typings/src/llm-providers/openai/OpenAiAssistantExecutionTools.d.ts +1 -1
- package/esm/typings/src/llm-providers/openai/register-configuration.d.ts +2 -2
- package/esm/typings/src/llm-providers/openai/register-constructor.d.ts +2 -2
- package/esm/typings/src/remote-server/socket-types/listModels/PromptbookServer_ListModels_Request.d.ts +1 -1
- package/esm/typings/src/scrapers/_boilerplate/createBoilerplateScraper.d.ts +1 -1
- package/esm/typings/src/scrapers/_boilerplate/register-constructor.d.ts +1 -1
- package/esm/typings/src/scrapers/_boilerplate/register-metadata.d.ts +2 -2
- package/esm/typings/src/scrapers/_common/prepareKnowledgePieces.d.ts +1 -1
- package/esm/typings/src/scrapers/_common/register/ScraperAndConverterMetadata.d.ts +1 -1
- package/esm/typings/src/scrapers/document/createDocumentScraper.d.ts +1 -1
- package/esm/typings/src/scrapers/document/register-constructor.d.ts +1 -1
- package/esm/typings/src/scrapers/document/register-metadata.d.ts +2 -2
- package/esm/typings/src/scrapers/document-legacy/createLegacyDocumentScraper.d.ts +1 -1
- package/esm/typings/src/scrapers/document-legacy/register-constructor.d.ts +1 -1
- package/esm/typings/src/scrapers/document-legacy/register-metadata.d.ts +2 -2
- package/esm/typings/src/scrapers/markdown/createMarkdownScraper.d.ts +1 -4
- package/esm/typings/src/scrapers/markdown/register-constructor.d.ts +1 -1
- package/esm/typings/src/scrapers/markdown/register-metadata.d.ts +2 -2
- package/esm/typings/src/scrapers/markitdown/createMarkitdownScraper.d.ts +1 -1
- package/esm/typings/src/scrapers/markitdown/register-constructor.d.ts +1 -1
- package/esm/typings/src/scrapers/markitdown/register-metadata.d.ts +2 -2
- package/esm/typings/src/scrapers/pdf/createPdfScraper.d.ts +1 -1
- package/esm/typings/src/scrapers/pdf/register-constructor.d.ts +1 -1
- package/esm/typings/src/scrapers/pdf/register-metadata.d.ts +2 -2
- package/esm/typings/src/scrapers/website/createWebsiteScraper.d.ts +1 -1
- package/esm/typings/src/scrapers/website/register-constructor.d.ts +1 -1
- package/esm/typings/src/scrapers/website/register-metadata.d.ts +2 -2
- package/esm/typings/src/types/typeAliases.d.ts +1 -1
- package/esm/typings/src/utils/files/listAllFiles.d.ts +1 -1
- package/esm/typings/src/version.d.ts +1 -1
- package/esm/typings/src/{wizzard → wizard}/$getCompiledBook.d.ts +2 -2
- package/esm/typings/src/{wizzard/wizzard.d.ts → wizard/wizard.d.ts} +6 -6
- package/package.json +25 -14
- package/umd/index.umd.js +192 -64
- package/umd/index.umd.js.map +1 -1
package/esm/index.es.js
CHANGED
|
@@ -18,12 +18,52 @@ const BOOK_LANGUAGE_VERSION = '1.0.0';
|
|
|
18
18
|
* @generated
|
|
19
19
|
* @see https://github.com/webgptorg/promptbook
|
|
20
20
|
*/
|
|
21
|
-
const PROMPTBOOK_ENGINE_VERSION = '0.
|
|
21
|
+
const PROMPTBOOK_ENGINE_VERSION = '0.95.0';
|
|
22
22
|
/**
|
|
23
23
|
* TODO: string_promptbook_version should be constrained to the all versions of Promptbook engine
|
|
24
24
|
* Note: [💞] Ignore a discrepancy between file name and entity name
|
|
25
25
|
*/
|
|
26
26
|
|
|
27
|
+
/**
|
|
28
|
+
* Detects if the code is running in a browser environment in main thread (Not in a web worker)
|
|
29
|
+
*
|
|
30
|
+
* Note: `$` is used to indicate that this function is not a pure function - it looks at the global object to determine the environment
|
|
31
|
+
*
|
|
32
|
+
* @public exported from `@promptbook/utils`
|
|
33
|
+
*/
|
|
34
|
+
const $isRunningInBrowser = new Function(`
|
|
35
|
+
try {
|
|
36
|
+
return this === window;
|
|
37
|
+
} catch (e) {
|
|
38
|
+
return false;
|
|
39
|
+
}
|
|
40
|
+
`);
|
|
41
|
+
/**
|
|
42
|
+
* TODO: [🎺]
|
|
43
|
+
*/
|
|
44
|
+
|
|
45
|
+
/**
|
|
46
|
+
* Detects if the code is running in a web worker
|
|
47
|
+
*
|
|
48
|
+
* Note: `$` is used to indicate that this function is not a pure function - it looks at the global object to determine the environment
|
|
49
|
+
*
|
|
50
|
+
* @public exported from `@promptbook/utils`
|
|
51
|
+
*/
|
|
52
|
+
const $isRunningInWebWorker = new Function(`
|
|
53
|
+
try {
|
|
54
|
+
if (typeof WorkerGlobalScope !== 'undefined' && self instanceof WorkerGlobalScope) {
|
|
55
|
+
return true;
|
|
56
|
+
} else {
|
|
57
|
+
return false;
|
|
58
|
+
}
|
|
59
|
+
} catch (e) {
|
|
60
|
+
return false;
|
|
61
|
+
}
|
|
62
|
+
`);
|
|
63
|
+
/**
|
|
64
|
+
* TODO: [🎺]
|
|
65
|
+
*/
|
|
66
|
+
|
|
27
67
|
/**
|
|
28
68
|
* Freezes the given object and all its nested objects recursively
|
|
29
69
|
*
|
|
@@ -405,7 +445,7 @@ function checkSerializableAsJson(options) {
|
|
|
405
445
|
else {
|
|
406
446
|
for (const [subName, subValue] of Object.entries(value)) {
|
|
407
447
|
if (subValue === undefined) {
|
|
408
|
-
// Note: undefined in object is serializable - it is just
|
|
448
|
+
// Note: undefined in object is serializable - it is just omitted
|
|
409
449
|
continue;
|
|
410
450
|
}
|
|
411
451
|
checkSerializableAsJson({ name: `${name}.${subName}`, value: subValue, message });
|
|
@@ -1055,7 +1095,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1055
1095
|
modelVariant: 'COMPLETION',
|
|
1056
1096
|
modelTitle: 'davinci-002',
|
|
1057
1097
|
modelName: 'davinci-002',
|
|
1058
|
-
modelDescription: 'Legacy completion model with
|
|
1098
|
+
modelDescription: 'Legacy completion model with 4K token context window. Excels at complex text generation, creative writing, and detailed content creation with strong contextual understanding. Optimized for instructions requiring nuanced outputs and extended reasoning. Suitable for applications needing high-quality text generation without conversation management.',
|
|
1059
1099
|
pricing: {
|
|
1060
1100
|
prompt: pricing(`$2.00 / 1M tokens`),
|
|
1061
1101
|
output: pricing(`$2.00 / 1M tokens`),
|
|
@@ -1073,7 +1113,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1073
1113
|
modelVariant: 'CHAT',
|
|
1074
1114
|
modelTitle: 'gpt-3.5-turbo-16k',
|
|
1075
1115
|
modelName: 'gpt-3.5-turbo-16k',
|
|
1076
|
-
modelDescription: 'GPT-3.5 Turbo with
|
|
1116
|
+
modelDescription: 'Extended context GPT-3.5 Turbo with 16K token window. Maintains core capabilities of standard 3.5 Turbo while supporting longer conversations and documents. Features good balance of performance and cost for applications requiring more context than standard 4K models. Effective for document analysis, extended conversations, and multi-step reasoning tasks.',
|
|
1077
1117
|
pricing: {
|
|
1078
1118
|
prompt: pricing(`$3.00 / 1M tokens`),
|
|
1079
1119
|
output: pricing(`$4.00 / 1M tokens`),
|
|
@@ -1097,7 +1137,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1097
1137
|
modelVariant: 'CHAT',
|
|
1098
1138
|
modelTitle: 'gpt-4',
|
|
1099
1139
|
modelName: 'gpt-4',
|
|
1100
|
-
modelDescription: '
|
|
1140
|
+
modelDescription: 'Powerful language model with 8K context window featuring sophisticated reasoning, instruction-following, and knowledge capabilities. Demonstrates strong performance on complex tasks requiring deep understanding and multi-step reasoning. Excels at code generation, logical analysis, and nuanced content creation. Suitable for advanced applications requiring high-quality outputs.',
|
|
1101
1141
|
pricing: {
|
|
1102
1142
|
prompt: pricing(`$30.00 / 1M tokens`),
|
|
1103
1143
|
output: pricing(`$60.00 / 1M tokens`),
|
|
@@ -1109,7 +1149,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1109
1149
|
modelVariant: 'CHAT',
|
|
1110
1150
|
modelTitle: 'gpt-4-32k',
|
|
1111
1151
|
modelName: 'gpt-4-32k',
|
|
1112
|
-
modelDescription: 'Extended context version of GPT-4 with
|
|
1152
|
+
modelDescription: 'Extended context version of GPT-4 with 32K token window. Maintains all capabilities of standard GPT-4 while supporting analysis of very lengthy documents, code bases, and conversations. Features enhanced ability to maintain context over long interactions and process detailed information from large inputs. Ideal for document analysis, legal review, and complex problem-solving.',
|
|
1113
1153
|
pricing: {
|
|
1114
1154
|
prompt: pricing(`$60.00 / 1M tokens`),
|
|
1115
1155
|
output: pricing(`$120.00 / 1M tokens`),
|
|
@@ -1132,7 +1172,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1132
1172
|
modelVariant: 'CHAT',
|
|
1133
1173
|
modelTitle: 'gpt-4-turbo-2024-04-09',
|
|
1134
1174
|
modelName: 'gpt-4-turbo-2024-04-09',
|
|
1135
|
-
modelDescription: 'Latest stable GPT-4 Turbo
|
|
1175
|
+
modelDescription: 'Latest stable GPT-4 Turbo from April 2024 with 128K context window. Features enhanced reasoning chains, improved factual accuracy with 40% reduction in hallucinations, and better instruction following compared to earlier versions. Includes advanced function calling capabilities and knowledge up to April 2024. Provides optimal performance for enterprise applications requiring reliability.',
|
|
1136
1176
|
pricing: {
|
|
1137
1177
|
prompt: pricing(`$10.00 / 1M tokens`),
|
|
1138
1178
|
output: pricing(`$30.00 / 1M tokens`),
|
|
@@ -1144,7 +1184,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1144
1184
|
modelVariant: 'CHAT',
|
|
1145
1185
|
modelTitle: 'gpt-3.5-turbo-1106',
|
|
1146
1186
|
modelName: 'gpt-3.5-turbo-1106',
|
|
1147
|
-
modelDescription: 'November 2023 version of GPT-3.5 Turbo with improved instruction following and
|
|
1187
|
+
modelDescription: 'November 2023 version of GPT-3.5 Turbo with 16K token context window. Features improved instruction following, more consistent output formatting, and enhanced function calling capabilities. Includes knowledge cutoff from April 2023. Suitable for applications requiring good performance at lower cost than GPT-4 models.',
|
|
1148
1188
|
pricing: {
|
|
1149
1189
|
prompt: pricing(`$1.00 / 1M tokens`),
|
|
1150
1190
|
output: pricing(`$2.00 / 1M tokens`),
|
|
@@ -1156,7 +1196,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1156
1196
|
modelVariant: 'CHAT',
|
|
1157
1197
|
modelTitle: 'gpt-4-turbo',
|
|
1158
1198
|
modelName: 'gpt-4-turbo',
|
|
1159
|
-
modelDescription: 'More capable
|
|
1199
|
+
modelDescription: 'More capable and cost-efficient version of GPT-4 with 128K token context window. Features improved instruction following, advanced function calling capabilities, and better performance on coding tasks. Maintains superior reasoning and knowledge while offering substantial cost reduction compared to base GPT-4. Ideal for complex applications requiring extensive context processing.',
|
|
1160
1200
|
pricing: {
|
|
1161
1201
|
prompt: pricing(`$10.00 / 1M tokens`),
|
|
1162
1202
|
output: pricing(`$30.00 / 1M tokens`),
|
|
@@ -1168,10 +1208,10 @@ const OPENAI_MODELS = exportJson({
|
|
|
1168
1208
|
modelVariant: 'COMPLETION',
|
|
1169
1209
|
modelTitle: 'gpt-3.5-turbo-instruct-0914',
|
|
1170
1210
|
modelName: 'gpt-3.5-turbo-instruct-0914',
|
|
1171
|
-
modelDescription: 'September 2023 version of GPT-3.5 Turbo
|
|
1211
|
+
modelDescription: 'September 2023 version of GPT-3.5 Turbo Instruct with 4K context window. Optimized for completion-style instruction following with deterministic responses. Better suited than chat models for applications requiring specific formatted outputs without conversation management. Knowledge cutoff from September 2021.',
|
|
1172
1212
|
pricing: {
|
|
1173
1213
|
prompt: pricing(`$1.50 / 1M tokens`),
|
|
1174
|
-
output: pricing(`$2.00 / 1M tokens`),
|
|
1214
|
+
output: pricing(`$2.00 / 1M tokens`),
|
|
1175
1215
|
},
|
|
1176
1216
|
},
|
|
1177
1217
|
/**/
|
|
@@ -1180,7 +1220,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1180
1220
|
modelVariant: 'COMPLETION',
|
|
1181
1221
|
modelTitle: 'gpt-3.5-turbo-instruct',
|
|
1182
1222
|
modelName: 'gpt-3.5-turbo-instruct',
|
|
1183
|
-
modelDescription: 'Optimized version of GPT-3.5 for completion-style API with
|
|
1223
|
+
modelDescription: 'Optimized version of GPT-3.5 for completion-style API with 4K token context window. Features strong instruction following with single-turn design rather than multi-turn conversation. Provides more consistent, deterministic outputs compared to chat models. Well-suited for templated content generation and structured text transformation tasks.',
|
|
1184
1224
|
pricing: {
|
|
1185
1225
|
prompt: pricing(`$1.50 / 1M tokens`),
|
|
1186
1226
|
output: pricing(`$2.00 / 1M tokens`),
|
|
@@ -1198,7 +1238,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1198
1238
|
modelVariant: 'CHAT',
|
|
1199
1239
|
modelTitle: 'gpt-3.5-turbo',
|
|
1200
1240
|
modelName: 'gpt-3.5-turbo',
|
|
1201
|
-
modelDescription: 'Latest version of GPT-3.5 Turbo with improved performance
|
|
1241
|
+
modelDescription: 'Latest version of GPT-3.5 Turbo with 4K token default context window (16K available). Features continually improved performance with enhanced instruction following and reduced hallucinations. Offers excellent balance between capability and cost efficiency. Suitable for most general-purpose applications requiring good AI capabilities at reasonable cost.',
|
|
1202
1242
|
pricing: {
|
|
1203
1243
|
prompt: pricing(`$0.50 / 1M tokens`),
|
|
1204
1244
|
output: pricing(`$1.50 / 1M tokens`),
|
|
@@ -1210,7 +1250,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1210
1250
|
modelVariant: 'CHAT',
|
|
1211
1251
|
modelTitle: 'gpt-3.5-turbo-0301',
|
|
1212
1252
|
modelName: 'gpt-3.5-turbo-0301',
|
|
1213
|
-
modelDescription: 'March 2023 version of GPT-3.5 Turbo with
|
|
1253
|
+
modelDescription: 'March 2023 version of GPT-3.5 Turbo with 4K token context window. Legacy model maintained for backward compatibility with specific application behaviors. Features solid conversational abilities and basic instruction following. Knowledge cutoff from September 2021. Suitable for applications explicitly designed for this version.',
|
|
1214
1254
|
pricing: {
|
|
1215
1255
|
prompt: pricing(`$1.50 / 1M tokens`),
|
|
1216
1256
|
output: pricing(`$2.00 / 1M tokens`),
|
|
@@ -1222,7 +1262,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1222
1262
|
modelVariant: 'COMPLETION',
|
|
1223
1263
|
modelTitle: 'babbage-002',
|
|
1224
1264
|
modelName: 'babbage-002',
|
|
1225
|
-
modelDescription: 'Efficient legacy completion model with
|
|
1265
|
+
modelDescription: 'Efficient legacy completion model with 4K context window balancing performance and speed. Features moderate reasoning capabilities with focus on straightforward text generation tasks. Significantly more efficient than davinci models while maintaining adequate quality for many applications. Suitable for high-volume, cost-sensitive text generation needs.',
|
|
1226
1266
|
pricing: {
|
|
1227
1267
|
prompt: pricing(`$0.40 / 1M tokens`),
|
|
1228
1268
|
output: pricing(`$0.40 / 1M tokens`),
|
|
@@ -1234,7 +1274,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1234
1274
|
modelVariant: 'CHAT',
|
|
1235
1275
|
modelTitle: 'gpt-4-1106-preview',
|
|
1236
1276
|
modelName: 'gpt-4-1106-preview',
|
|
1237
|
-
modelDescription: 'November 2023 preview version of GPT-4 Turbo with improved instruction following and
|
|
1277
|
+
modelDescription: 'November 2023 preview version of GPT-4 Turbo with 128K token context window. Features improved instruction following, better function calling capabilities, and enhanced reasoning. Includes knowledge cutoff from April 2023. Suitable for complex applications requiring extensive document understanding and sophisticated interactions.',
|
|
1238
1278
|
pricing: {
|
|
1239
1279
|
prompt: pricing(`$10.00 / 1M tokens`),
|
|
1240
1280
|
output: pricing(`$30.00 / 1M tokens`),
|
|
@@ -1246,7 +1286,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1246
1286
|
modelVariant: 'CHAT',
|
|
1247
1287
|
modelTitle: 'gpt-4-0125-preview',
|
|
1248
1288
|
modelName: 'gpt-4-0125-preview',
|
|
1249
|
-
modelDescription: 'January 2024 preview version of GPT-4 Turbo with improved reasoning capabilities and
|
|
1289
|
+
modelDescription: 'January 2024 preview version of GPT-4 Turbo with 128K token context window. Features improved reasoning capabilities, enhanced tool use, and more reliable function calling. Includes knowledge cutoff from October 2023. Offers better performance on complex logical tasks and more consistent outputs than previous preview versions.',
|
|
1250
1290
|
pricing: {
|
|
1251
1291
|
prompt: pricing(`$10.00 / 1M tokens`),
|
|
1252
1292
|
output: pricing(`$30.00 / 1M tokens`),
|
|
@@ -1264,7 +1304,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1264
1304
|
modelVariant: 'CHAT',
|
|
1265
1305
|
modelTitle: 'gpt-3.5-turbo-0125',
|
|
1266
1306
|
modelName: 'gpt-3.5-turbo-0125',
|
|
1267
|
-
modelDescription: 'January 2024 version of GPT-3.5 Turbo with improved reasoning capabilities and
|
|
1307
|
+
modelDescription: 'January 2024 version of GPT-3.5 Turbo with 16K token context window. Features improved reasoning capabilities, better instruction adherence, and reduced hallucinations compared to previous versions. Includes knowledge cutoff from September 2021. Provides good performance for most general applications at reasonable cost.',
|
|
1268
1308
|
pricing: {
|
|
1269
1309
|
prompt: pricing(`$0.50 / 1M tokens`),
|
|
1270
1310
|
output: pricing(`$1.50 / 1M tokens`),
|
|
@@ -1276,7 +1316,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1276
1316
|
modelVariant: 'CHAT',
|
|
1277
1317
|
modelTitle: 'gpt-4-turbo-preview',
|
|
1278
1318
|
modelName: 'gpt-4-turbo-preview',
|
|
1279
|
-
modelDescription: 'Preview version of GPT-4 Turbo that points to the latest model
|
|
1319
|
+
modelDescription: 'Preview version of GPT-4 Turbo with 128K token context window that points to the latest development model. Features cutting-edge improvements to instruction following, knowledge representation, and tool use capabilities. Provides access to newest features but may have occasional behavior changes. Best for non-critical applications wanting latest capabilities.',
|
|
1280
1320
|
pricing: {
|
|
1281
1321
|
prompt: pricing(`$10.00 / 1M tokens`),
|
|
1282
1322
|
output: pricing(`$30.00 / 1M tokens`),
|
|
@@ -1288,11 +1328,10 @@ const OPENAI_MODELS = exportJson({
|
|
|
1288
1328
|
modelVariant: 'EMBEDDING',
|
|
1289
1329
|
modelTitle: 'text-embedding-3-large',
|
|
1290
1330
|
modelName: 'text-embedding-3-large',
|
|
1291
|
-
modelDescription: "OpenAI's most capable text embedding model
|
|
1331
|
+
modelDescription: "OpenAI's most capable text embedding model generating 3072-dimensional vectors. Designed for high-quality embeddings for complex similarity tasks, clustering, and information retrieval. Features enhanced cross-lingual capabilities and significantly improved performance on retrieval and classification benchmarks. Ideal for sophisticated RAG systems and semantic search applications.",
|
|
1292
1332
|
pricing: {
|
|
1293
1333
|
prompt: pricing(`$0.13 / 1M tokens`),
|
|
1294
|
-
|
|
1295
|
-
output: 0, // <- Note: [🆖] In Embedding models you dont pay for output
|
|
1334
|
+
output: 0,
|
|
1296
1335
|
},
|
|
1297
1336
|
},
|
|
1298
1337
|
/**/
|
|
@@ -1301,11 +1340,10 @@ const OPENAI_MODELS = exportJson({
|
|
|
1301
1340
|
modelVariant: 'EMBEDDING',
|
|
1302
1341
|
modelTitle: 'text-embedding-3-small',
|
|
1303
1342
|
modelName: 'text-embedding-3-small',
|
|
1304
|
-
modelDescription: 'Cost-effective embedding model
|
|
1343
|
+
modelDescription: 'Cost-effective embedding model generating 1536-dimensional vectors. Balances quality and efficiency for simpler tasks while maintaining good performance on text similarity and retrieval applications. Offers 20% better quality than ada-002 at significantly lower cost. Ideal for production embedding applications with cost constraints.',
|
|
1305
1344
|
pricing: {
|
|
1306
1345
|
prompt: pricing(`$0.02 / 1M tokens`),
|
|
1307
|
-
|
|
1308
|
-
output: 0, // <- Note: [🆖] In Embedding models you dont pay for output
|
|
1346
|
+
output: 0,
|
|
1309
1347
|
},
|
|
1310
1348
|
},
|
|
1311
1349
|
/**/
|
|
@@ -1314,7 +1352,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1314
1352
|
modelVariant: 'CHAT',
|
|
1315
1353
|
modelTitle: 'gpt-3.5-turbo-0613',
|
|
1316
1354
|
modelName: 'gpt-3.5-turbo-0613',
|
|
1317
|
-
modelDescription:
|
|
1355
|
+
modelDescription: "June 2023 version of GPT-3.5 Turbo with 4K token context window. Features function calling capabilities for structured data extraction and API interaction. Includes knowledge cutoff from September 2021. Maintained for applications specifically designed for this version's behaviors and capabilities.",
|
|
1318
1356
|
pricing: {
|
|
1319
1357
|
prompt: pricing(`$1.50 / 1M tokens`),
|
|
1320
1358
|
output: pricing(`$2.00 / 1M tokens`),
|
|
@@ -1326,11 +1364,10 @@ const OPENAI_MODELS = exportJson({
|
|
|
1326
1364
|
modelVariant: 'EMBEDDING',
|
|
1327
1365
|
modelTitle: 'text-embedding-ada-002',
|
|
1328
1366
|
modelName: 'text-embedding-ada-002',
|
|
1329
|
-
modelDescription: 'Legacy text embedding model suitable for text similarity and retrieval
|
|
1367
|
+
modelDescription: 'Legacy text embedding model generating 1536-dimensional vectors suitable for text similarity and retrieval applications. Processes up to 8K tokens per request with consistent embedding quality. While superseded by newer embedding-3 models, still maintains adequate performance for many semantic search and classification tasks.',
|
|
1330
1368
|
pricing: {
|
|
1331
1369
|
prompt: pricing(`$0.1 / 1M tokens`),
|
|
1332
|
-
|
|
1333
|
-
output: 0, // <- Note: [🆖] In Embedding models you dont pay for output
|
|
1370
|
+
output: 0,
|
|
1334
1371
|
},
|
|
1335
1372
|
},
|
|
1336
1373
|
/**/
|
|
@@ -1357,7 +1394,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1357
1394
|
modelVariant: 'CHAT',
|
|
1358
1395
|
modelTitle: 'gpt-4o-2024-05-13',
|
|
1359
1396
|
modelName: 'gpt-4o-2024-05-13',
|
|
1360
|
-
modelDescription: 'May 2024 version of GPT-4o with enhanced multimodal capabilities
|
|
1397
|
+
modelDescription: 'May 2024 version of GPT-4o with 128K context window. Features enhanced multimodal capabilities including superior image understanding (up to 20MP), audio processing, and improved reasoning. Optimized for 2x lower latency than GPT-4 Turbo while maintaining high performance. Includes knowledge up to October 2023. Ideal for production applications requiring reliable multimodal capabilities.',
|
|
1361
1398
|
pricing: {
|
|
1362
1399
|
prompt: pricing(`$5.00 / 1M tokens`),
|
|
1363
1400
|
output: pricing(`$15.00 / 1M tokens`),
|
|
@@ -1369,7 +1406,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1369
1406
|
modelVariant: 'CHAT',
|
|
1370
1407
|
modelTitle: 'gpt-4o',
|
|
1371
1408
|
modelName: 'gpt-4o',
|
|
1372
|
-
modelDescription: "OpenAI's most advanced multimodal model
|
|
1409
|
+
modelDescription: "OpenAI's most advanced general-purpose multimodal model with 128K context window. Optimized for balanced performance, speed, and cost with 2x faster responses than GPT-4 Turbo. Features excellent vision processing, audio understanding, reasoning, and text generation quality. Represents optimal balance of capability and efficiency for most advanced applications.",
|
|
1373
1410
|
pricing: {
|
|
1374
1411
|
prompt: pricing(`$5.00 / 1M tokens`),
|
|
1375
1412
|
output: pricing(`$15.00 / 1M tokens`),
|
|
@@ -1381,7 +1418,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1381
1418
|
modelVariant: 'CHAT',
|
|
1382
1419
|
modelTitle: 'gpt-4o-mini',
|
|
1383
1420
|
modelName: 'gpt-4o-mini',
|
|
1384
|
-
modelDescription: 'Smaller, more cost-effective version of GPT-4o with
|
|
1421
|
+
modelDescription: 'Smaller, more cost-effective version of GPT-4o with 128K context window. Maintains impressive capabilities across text, vision, and audio tasks while operating at significantly lower cost. Features 3x faster inference than GPT-4o with good performance on general tasks. Excellent for applications requiring good quality multimodal capabilities at scale.',
|
|
1385
1422
|
pricing: {
|
|
1386
1423
|
prompt: pricing(`$0.15 / 1M tokens`),
|
|
1387
1424
|
output: pricing(`$0.60 / 1M tokens`),
|
|
@@ -1393,7 +1430,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1393
1430
|
modelVariant: 'CHAT',
|
|
1394
1431
|
modelTitle: 'o1-preview',
|
|
1395
1432
|
modelName: 'o1-preview',
|
|
1396
|
-
modelDescription: 'Advanced reasoning model with
|
|
1433
|
+
modelDescription: 'Advanced reasoning model with 128K context window specializing in complex logical, mathematical, and analytical tasks. Features exceptional step-by-step problem-solving capabilities, advanced mathematical and scientific reasoning, and superior performance on STEM-focused problems. Significantly outperforms GPT-4 on quantitative reasoning benchmarks. Ideal for professional and specialized applications.',
|
|
1397
1434
|
pricing: {
|
|
1398
1435
|
prompt: pricing(`$15.00 / 1M tokens`),
|
|
1399
1436
|
output: pricing(`$60.00 / 1M tokens`),
|
|
@@ -1405,8 +1442,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1405
1442
|
modelVariant: 'CHAT',
|
|
1406
1443
|
modelTitle: 'o1-preview-2024-09-12',
|
|
1407
1444
|
modelName: 'o1-preview-2024-09-12',
|
|
1408
|
-
modelDescription: 'September 2024 version of O1 preview with specialized reasoning capabilities for
|
|
1409
|
-
// <- TODO: [💩] Some better system to organize these date suffixes and versions
|
|
1445
|
+
modelDescription: 'September 2024 version of O1 preview with 128K context window. Features specialized reasoning capabilities with 30% improvement on mathematical and scientific accuracy over previous versions. Includes enhanced support for formal logic, statistical analysis, and technical domains. Optimized for professional applications requiring precise analytical thinking and rigorous methodologies.',
|
|
1410
1446
|
pricing: {
|
|
1411
1447
|
prompt: pricing(`$15.00 / 1M tokens`),
|
|
1412
1448
|
output: pricing(`$60.00 / 1M tokens`),
|
|
@@ -1418,7 +1454,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1418
1454
|
modelVariant: 'CHAT',
|
|
1419
1455
|
modelTitle: 'o1-mini',
|
|
1420
1456
|
modelName: 'o1-mini',
|
|
1421
|
-
modelDescription: 'Smaller, cost-effective version of the O1 model with good performance on
|
|
1457
|
+
modelDescription: 'Smaller, cost-effective version of the O1 model with 128K context window. Maintains strong analytical reasoning abilities while reducing computational requirements by 70%. Features good performance on mathematical, logical, and scientific tasks at significantly lower cost than full O1. Excellent for everyday analytical applications that benefit from reasoning focus.',
|
|
1422
1458
|
pricing: {
|
|
1423
1459
|
prompt: pricing(`$3.00 / 1M tokens`),
|
|
1424
1460
|
output: pricing(`$12.00 / 1M tokens`),
|
|
@@ -1430,7 +1466,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1430
1466
|
modelVariant: 'CHAT',
|
|
1431
1467
|
modelTitle: 'o1',
|
|
1432
1468
|
modelName: 'o1',
|
|
1433
|
-
modelDescription: "OpenAI's advanced reasoning model
|
|
1469
|
+
modelDescription: "OpenAI's advanced reasoning model with 128K context window focusing on logical problem-solving and analytical thinking. Features exceptional performance on quantitative tasks, step-by-step deduction, and complex technical problems. Maintains 95%+ of o1-preview capabilities with production-ready stability. Ideal for scientific computing, financial analysis, and professional applications.",
|
|
1434
1470
|
pricing: {
|
|
1435
1471
|
prompt: pricing(`$15.00 / 1M tokens`),
|
|
1436
1472
|
output: pricing(`$60.00 / 1M tokens`),
|
|
@@ -1442,11 +1478,10 @@ const OPENAI_MODELS = exportJson({
|
|
|
1442
1478
|
modelVariant: 'CHAT',
|
|
1443
1479
|
modelTitle: 'o3-mini',
|
|
1444
1480
|
modelName: 'o3-mini',
|
|
1445
|
-
modelDescription: 'Cost-effective reasoning model optimized for academic and scientific problem-solving.
|
|
1481
|
+
modelDescription: 'Cost-effective reasoning model with 128K context window optimized for academic and scientific problem-solving. Features efficient performance on STEM tasks with specialized capabilities in mathematics, physics, chemistry, and computer science. Offers 80% of O1 performance on technical domains at significantly lower cost. Ideal for educational applications and research support.',
|
|
1446
1482
|
pricing: {
|
|
1447
1483
|
prompt: pricing(`$3.00 / 1M tokens`),
|
|
1448
1484
|
output: pricing(`$12.00 / 1M tokens`),
|
|
1449
|
-
// <- TODO: !! Unsure, check the pricing
|
|
1450
1485
|
},
|
|
1451
1486
|
},
|
|
1452
1487
|
/**/
|
|
@@ -1455,7 +1490,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1455
1490
|
modelVariant: 'CHAT',
|
|
1456
1491
|
modelTitle: 'o1-mini-2024-09-12',
|
|
1457
1492
|
modelName: 'o1-mini-2024-09-12',
|
|
1458
|
-
modelDescription: "September 2024 version of O1-mini with balanced reasoning capabilities and cost-efficiency.
|
|
1493
|
+
modelDescription: "September 2024 version of O1-mini with 128K context window featuring balanced reasoning capabilities and cost-efficiency. Includes 25% improvement in mathematical accuracy and enhanced performance on coding tasks compared to previous versions. Maintains efficient resource utilization while delivering improved results for analytical applications that don't require the full O1 model.",
|
|
1459
1494
|
pricing: {
|
|
1460
1495
|
prompt: pricing(`$3.00 / 1M tokens`),
|
|
1461
1496
|
output: pricing(`$12.00 / 1M tokens`),
|
|
@@ -1467,7 +1502,7 @@ const OPENAI_MODELS = exportJson({
|
|
|
1467
1502
|
modelVariant: 'CHAT',
|
|
1468
1503
|
modelTitle: 'gpt-3.5-turbo-16k-0613',
|
|
1469
1504
|
modelName: 'gpt-3.5-turbo-16k-0613',
|
|
1470
|
-
modelDescription:
|
|
1505
|
+
modelDescription: "June 2023 version of GPT-3.5 Turbo with extended 16K token context window. Features good handling of longer conversations and documents with improved memory management across extended contexts. Includes knowledge cutoff from September 2021. Maintained for applications specifically designed for this version's behaviors and capabilities.",
|
|
1471
1506
|
pricing: {
|
|
1472
1507
|
prompt: pricing(`$3.00 / 1M tokens`),
|
|
1473
1508
|
output: pricing(`$4.00 / 1M tokens`),
|
|
@@ -2056,131 +2091,221 @@ class OpenAiCompatibleExecutionTools {
|
|
|
2056
2091
|
const OLLAMA_MODELS = exportJson({
|
|
2057
2092
|
name: 'OLLAMA_MODELS',
|
|
2058
2093
|
value: [
|
|
2094
|
+
{
|
|
2095
|
+
modelVariant: 'CHAT',
|
|
2096
|
+
modelTitle: 'llama3',
|
|
2097
|
+
modelName: 'llama3',
|
|
2098
|
+
modelDescription: 'Meta Llama 3 (8B-70B parameters) with 8K context window. Latest generation foundation model with enhanced reasoning, instruction following, and factual accuracy. Superior performance to Llama 2 across all benchmarks with improved multilingual capabilities.',
|
|
2099
|
+
},
|
|
2100
|
+
{
|
|
2101
|
+
modelVariant: 'CHAT',
|
|
2102
|
+
modelTitle: 'llama3-chat',
|
|
2103
|
+
modelName: 'llama3-chat',
|
|
2104
|
+
modelDescription: 'Meta Llama 3 Chat with 8K context window, fine-tuned specifically for dialogue with significantly improved instruction following. Features enhanced safety guardrails and reduced hallucinations. Recommended over Llama 2 Chat for all conversational applications.',
|
|
2105
|
+
},
|
|
2106
|
+
{
|
|
2107
|
+
modelVariant: 'CHAT',
|
|
2108
|
+
modelTitle: 'llama3-instruct',
|
|
2109
|
+
modelName: 'llama3-instruct',
|
|
2110
|
+
modelDescription: 'Meta Llama 3 Instruct with 8K context window, fine-tuned for following specific instructions with precise outputs. Features improved structured response formatting and accurate completion of complex directives.',
|
|
2111
|
+
},
|
|
2112
|
+
{
|
|
2113
|
+
modelVariant: 'CHAT',
|
|
2114
|
+
modelTitle: 'codellama:13b',
|
|
2115
|
+
modelName: 'codellama:13b',
|
|
2116
|
+
modelDescription: 'Meta CodeLlama 13B with 16K context window, specialized foundation model for code generation and understanding. Supports multiple programming languages with strong contextual code completion capabilities.',
|
|
2117
|
+
},
|
|
2118
|
+
{
|
|
2119
|
+
modelVariant: 'CHAT',
|
|
2120
|
+
modelTitle: 'codellama:34b',
|
|
2121
|
+
modelName: 'codellama:34b',
|
|
2122
|
+
modelDescription: 'Meta CodeLlama 34B with 16K context window, larger code-specialized model with improved reasoning about complex programming tasks. Enhanced documentation generation and bug detection compared to smaller variants.',
|
|
2123
|
+
},
|
|
2124
|
+
{
|
|
2125
|
+
modelVariant: 'CHAT',
|
|
2126
|
+
modelTitle: 'phi3:mini',
|
|
2127
|
+
modelName: 'phi3:mini',
|
|
2128
|
+
modelDescription: 'Microsoft Phi-3 Mini (3.8B parameters) with 4K context window, highly efficient small language model with remarkable reasoning given its size. Performs competitively with much larger models on common benchmarks. Excellent for resource-constrained environments.',
|
|
2129
|
+
},
|
|
2130
|
+
{
|
|
2131
|
+
modelVariant: 'CHAT',
|
|
2132
|
+
modelTitle: 'phi3:medium',
|
|
2133
|
+
modelName: 'phi3:medium',
|
|
2134
|
+
modelDescription: 'Microsoft Phi-3 Medium (14B parameters) with 8K context window, balanced model offering strong performance with reasonable compute requirements. Features improved reasoning and factuality compared to Mini variant while maintaining efficiency.',
|
|
2135
|
+
},
|
|
2136
|
+
{
|
|
2137
|
+
modelVariant: 'CHAT',
|
|
2138
|
+
modelTitle: 'mistral-nemo',
|
|
2139
|
+
modelName: 'mistral-nemo',
|
|
2140
|
+
modelDescription: 'Mistral NeMo with 32K context window, open-source model optimized for enterprise use cases with improved reasoning and knowledge capabilities. Features strong performance on professional and domain-specific tasks.',
|
|
2141
|
+
},
|
|
2059
2142
|
{
|
|
2060
2143
|
modelVariant: 'CHAT',
|
|
2061
2144
|
modelTitle: 'llama2',
|
|
2062
2145
|
modelName: 'llama2',
|
|
2063
|
-
modelDescription: 'Meta Llama 2
|
|
2146
|
+
modelDescription: 'Meta Llama 2 (7B-70B parameters) with 4K context window. General-purpose foundation model balancing performance and efficiency for text generation and reasoning tasks. Suitable for most non-specialized applications. Note: Superseded by Llama 3 models which offer better performance.',
|
|
2064
2147
|
},
|
|
2065
2148
|
{
|
|
2066
2149
|
modelVariant: 'CHAT',
|
|
2067
2150
|
modelTitle: 'llama2-chat',
|
|
2068
2151
|
modelName: 'llama2-chat',
|
|
2069
|
-
modelDescription: 'Meta Llama 2 Chat,
|
|
2152
|
+
modelDescription: 'Meta Llama 2 Chat with 4K context window, fine-tuned specifically for conversational AI. Enhanced instruction following and safer responses compared to base model. Ideal for chatbots and interactive applications. Note: Consider using newer Llama 3 Chat for improved performance.',
|
|
2070
2153
|
},
|
|
2071
2154
|
{
|
|
2072
2155
|
modelVariant: 'CHAT',
|
|
2073
2156
|
modelTitle: 'alpaca-7b',
|
|
2074
2157
|
modelName: 'alpaca-7b',
|
|
2075
|
-
modelDescription: 'Stanford Alpaca 7B, instruction-tuned LLaMA model.',
|
|
2158
|
+
modelDescription: 'Stanford Alpaca 7B with 2K context window, instruction-tuned LLaMA model focused on following specific directions. Optimized for resource efficiency while maintaining good response quality. Suitable for lightweight applications.',
|
|
2076
2159
|
},
|
|
2077
2160
|
{
|
|
2078
2161
|
modelVariant: 'CHAT',
|
|
2079
2162
|
modelTitle: 'alpaca-30b',
|
|
2080
2163
|
modelName: 'alpaca-30b',
|
|
2081
|
-
modelDescription: 'Stanford Alpaca 30B
|
|
2164
|
+
modelDescription: 'Stanford Alpaca 30B with 2K context window. Larger instruction-tuned LLaMA model with improved reasoning and content generation capabilities. Better performance than 7B variant but requires more computational resources.',
|
|
2082
2165
|
},
|
|
2083
2166
|
{
|
|
2084
2167
|
modelVariant: 'CHAT',
|
|
2085
2168
|
modelTitle: 'vicuna-13b',
|
|
2086
2169
|
modelName: 'vicuna-13b',
|
|
2087
|
-
modelDescription: 'Vicuna 13B, fine-tuned LLaMA for chat and instruction.',
|
|
2170
|
+
modelDescription: 'Vicuna 13B with 2K context window, fine-tuned from LLaMA for chat and instruction following. Known for balanced performance, good conversational abilities, and improved helpfulness over base models. Popular for diverse conversational applications.',
|
|
2088
2171
|
},
|
|
2089
2172
|
{
|
|
2090
2173
|
modelVariant: 'CHAT',
|
|
2091
2174
|
modelTitle: 'falcon-7b',
|
|
2092
2175
|
modelName: 'falcon-7b',
|
|
2093
|
-
modelDescription: 'Falcon 7B,
|
|
2176
|
+
modelDescription: 'Falcon 7B with 2K context window, performant open large language model trained on 1.5 trillion tokens. Strong on general knowledge tasks with smaller computational requirements. Good balance of performance and efficiency.',
|
|
2094
2177
|
},
|
|
2095
2178
|
{
|
|
2096
2179
|
modelVariant: 'CHAT',
|
|
2097
2180
|
modelTitle: 'falcon-40b',
|
|
2098
2181
|
modelName: 'falcon-40b',
|
|
2099
|
-
modelDescription: 'Falcon 40B,
|
|
2182
|
+
modelDescription: 'Falcon 40B with 2K context window, larger open large language model with enhanced reasoning and knowledge capabilities. Significantly better performance than 7B version but requires substantially more resources. Suitable for complex generation tasks.',
|
|
2100
2183
|
},
|
|
2101
2184
|
{
|
|
2102
2185
|
modelVariant: 'CHAT',
|
|
2103
2186
|
modelTitle: 'bloom-7b',
|
|
2104
2187
|
modelName: 'bloom-7b',
|
|
2105
|
-
modelDescription: 'BLOOM 7B, multilingual large language model.',
|
|
2188
|
+
modelDescription: 'BLOOM 7B with 2K context window, multilingual large language model supporting 46+ languages. Trained for diverse linguistic capabilities across languages. Especially useful for non-English or multilingual applications.',
|
|
2106
2189
|
},
|
|
2107
2190
|
{
|
|
2108
2191
|
modelVariant: 'CHAT',
|
|
2109
2192
|
modelTitle: 'mistral-7b',
|
|
2110
2193
|
modelName: 'mistral-7b',
|
|
2111
|
-
modelDescription: 'Mistral 7B, efficient and fast open LLM.',
|
|
2194
|
+
modelDescription: 'Mistral 7B with 8K context window, efficient and fast open LLM with performance rivaling much larger models. Features Grouped-Query Attention for faster inference. Excellent balance of quality and speed for various applications.',
|
|
2112
2195
|
},
|
|
2113
2196
|
{
|
|
2114
2197
|
modelVariant: 'CHAT',
|
|
2115
2198
|
modelTitle: 'gorilla',
|
|
2116
2199
|
modelName: 'gorilla',
|
|
2117
|
-
modelDescription: 'Gorilla, open-source LLM for tool use and
|
|
2200
|
+
modelDescription: 'Gorilla with 4K context window, specialized open-source LLM for tool use and API interaction. Fine-tuned to understand and generate API calls accurately. Optimal for agent applications that interact with external tools and services.',
|
|
2118
2201
|
},
|
|
2119
2202
|
{
|
|
2120
2203
|
modelVariant: 'CHAT',
|
|
2121
2204
|
modelTitle: 'cerebras-13b',
|
|
2122
2205
|
modelName: 'cerebras-13b',
|
|
2123
|
-
modelDescription: 'Cerebras-GPT 13B,
|
|
2206
|
+
modelDescription: 'Cerebras-GPT 13B with 2K context window, trained on diverse high-quality datasets. Good general-purpose capabilities with particular strengths in factual response accuracy. Well-suited for applications requiring reliable information.',
|
|
2124
2207
|
},
|
|
2125
2208
|
{
|
|
2126
2209
|
modelVariant: 'CHAT',
|
|
2127
2210
|
modelTitle: 'openchat-7b',
|
|
2128
2211
|
modelName: 'openchat-7b',
|
|
2129
|
-
modelDescription: 'OpenChat 7B,
|
|
2212
|
+
modelDescription: 'OpenChat 7B with 4K context window, optimized for conversational abilities and instruction following. Outperforms many larger models on benchmark tasks while maintaining efficiency. Ideal for interactive applications with limited resources.',
|
|
2130
2213
|
},
|
|
2131
2214
|
{
|
|
2132
2215
|
modelVariant: 'CHAT',
|
|
2133
2216
|
modelTitle: 'openchat-13b',
|
|
2134
2217
|
modelName: 'openchat-13b',
|
|
2135
|
-
modelDescription: 'OpenChat 13B, larger conversational LLM.',
|
|
2218
|
+
modelDescription: 'OpenChat 13B with 4K context window, larger conversational LLM with enhanced reasoning, helpfulness, and knowledge. Significant improvement over 7B variant in complex tasks and nuanced conversations. Well-balanced for most conversational applications.',
|
|
2136
2219
|
},
|
|
2137
2220
|
{
|
|
2138
2221
|
modelVariant: 'CHAT',
|
|
2139
2222
|
modelTitle: 'mpt-7b-chat',
|
|
2140
2223
|
modelName: 'mpt-7b-chat',
|
|
2141
|
-
modelDescription: 'MPT-7B Chat, optimized for dialogue
|
|
2224
|
+
modelDescription: 'MPT-7B Chat with 4K context window (extendable to 65K+), optimized for dialogue using high-quality conversation data. Features enhanced conversational abilities with strong safety alignments. Good for deployment in public-facing chat applications.',
|
|
2142
2225
|
},
|
|
2143
2226
|
{
|
|
2144
2227
|
modelVariant: 'CHAT',
|
|
2145
2228
|
modelTitle: 'mpt-7b-instruct',
|
|
2146
2229
|
modelName: 'mpt-7b-instruct',
|
|
2147
|
-
modelDescription: 'MPT-7B Instruct, instruction-tuned variant.',
|
|
2230
|
+
modelDescription: 'MPT-7B Instruct with 4K context window (extendable to 65K+), instruction-tuned variant optimized for following specific directives. Better than chat variant for single-turn instruction tasks. Well-suited for content generation and task completion.',
|
|
2148
2231
|
},
|
|
2149
2232
|
{
|
|
2150
2233
|
modelVariant: 'CHAT',
|
|
2151
2234
|
modelTitle: 'command-7b',
|
|
2152
2235
|
modelName: 'command-7b',
|
|
2153
|
-
modelDescription: 'Command 7B, instruction-following LLM.',
|
|
2236
|
+
modelDescription: 'Command 7B with 4K context window, instruction-following LLM tuned specifically for direct command execution and helpful responses. Optimized for clarity of outputs and following explicit directions. Good for practical task-oriented applications.',
|
|
2154
2237
|
},
|
|
2155
2238
|
{
|
|
2156
2239
|
modelVariant: 'CHAT',
|
|
2157
2240
|
modelTitle: 'starcoder',
|
|
2158
2241
|
modelName: 'starcoder',
|
|
2159
|
-
modelDescription: 'StarCoder, code generation large language model.',
|
|
2242
|
+
modelDescription: 'StarCoder with 8K context window, specialized code generation large language model trained on permissively licensed code. Supports 80+ programming languages. Optimized for code completion, generation, and understanding tasks.',
|
|
2160
2243
|
},
|
|
2161
2244
|
{
|
|
2162
2245
|
modelVariant: 'CHAT',
|
|
2163
2246
|
modelTitle: 'starcoder2',
|
|
2164
2247
|
modelName: 'starcoder2',
|
|
2165
|
-
modelDescription: 'StarCoder2, improved code generation model.',
|
|
2248
|
+
modelDescription: 'StarCoder2 with 16K context window, improved code generation model with better reasoning about code, debugging capabilities, and documentation generation. Supports 600+ programming languages. Ideal for complex software development assistance.',
|
|
2166
2249
|
},
|
|
2167
2250
|
{
|
|
2168
2251
|
modelVariant: 'CHAT',
|
|
2169
2252
|
modelTitle: 'mixtral-7b-chat',
|
|
2170
2253
|
modelName: 'mixtral-7b-chat',
|
|
2171
|
-
modelDescription: 'Mixtral 7B Chat, Mixture-of-Experts conversational model.',
|
|
2254
|
+
modelDescription: 'Mixtral 7B Chat with 32K context window, Mixture-of-Experts conversational model with strong performance across diverse tasks. Efficiently routes inputs to specialized sub-networks for optimal processing. Well-balanced for most chat applications.',
|
|
2172
2255
|
},
|
|
2173
2256
|
{
|
|
2174
2257
|
modelVariant: 'CHAT',
|
|
2175
2258
|
modelTitle: 'mixtral-8x7b',
|
|
2176
2259
|
modelName: 'mixtral-8x7b',
|
|
2177
|
-
modelDescription: 'Mixtral 8x7B, Mixture-of-Experts
|
|
2260
|
+
modelDescription: 'Mixtral 8x7B with 32K context window, advanced Mixture-of-Experts architecture using 8 expert networks of 7B parameters each. Competitive with much larger dense models while using less computation. Excellent general-purpose capabilities.',
|
|
2178
2261
|
},
|
|
2179
2262
|
{
|
|
2180
2263
|
modelVariant: 'CHAT',
|
|
2181
2264
|
modelTitle: 'mixtral-8x7b-instruct',
|
|
2182
2265
|
modelName: 'mixtral-8x7b-instruct',
|
|
2183
|
-
modelDescription: 'Mixtral 8x7B Instruct, instruction-tuned Mixture-of-Experts model.',
|
|
2266
|
+
modelDescription: 'Mixtral 8x7B Instruct with 32K context window, instruction-tuned Mixture-of-Experts model for direct task execution. Enhanced directive following and more structured outputs compared to base model. Optimal for specific instruction-based workloads.',
|
|
2267
|
+
},
|
|
2268
|
+
{
|
|
2269
|
+
modelVariant: 'CHAT',
|
|
2270
|
+
modelTitle: 'neural-chat',
|
|
2271
|
+
modelName: 'neural-chat',
|
|
2272
|
+
modelDescription: 'Intel Neural Chat (latest) with 8K context window, optimized for Intel hardware with efficient inference. Balanced model for general-purpose conversational applications with good instruction following capabilities.',
|
|
2273
|
+
},
|
|
2274
|
+
{
|
|
2275
|
+
modelVariant: 'CHAT',
|
|
2276
|
+
modelTitle: 'qwen:7b',
|
|
2277
|
+
modelName: 'qwen:7b',
|
|
2278
|
+
modelDescription: 'Alibaba Qwen 7B with 8K context window. Versatile model with strong multilingual capabilities, particularly for Chinese and English. Features good reasoning and knowledge across diverse domains.',
|
|
2279
|
+
},
|
|
2280
|
+
{
|
|
2281
|
+
modelVariant: 'CHAT',
|
|
2282
|
+
modelTitle: 'qwen:14b',
|
|
2283
|
+
modelName: 'qwen:14b',
|
|
2284
|
+
modelDescription: 'Alibaba Qwen 14B with 8K context window. Enhanced version with improved reasoning and knowledge capabilities. Particularly strong in multilingual applications and domain-specific tasks requiring deeper understanding.',
|
|
2285
|
+
},
|
|
2286
|
+
{
|
|
2287
|
+
modelVariant: 'CHAT',
|
|
2288
|
+
modelTitle: 'gemma:2b',
|
|
2289
|
+
modelName: 'gemma:2b',
|
|
2290
|
+
modelDescription: 'Google Gemma 2B with 8K context window. Lightweight but capable model designed for efficiency. Good performance for its size on common tasks, ideal for resource-constrained environments and quick responses.',
|
|
2291
|
+
},
|
|
2292
|
+
{
|
|
2293
|
+
modelVariant: 'CHAT',
|
|
2294
|
+
modelTitle: 'gemma:7b',
|
|
2295
|
+
modelName: 'gemma:7b',
|
|
2296
|
+
modelDescription: 'Google Gemma 7B with 8K context window. Well-balanced model offering strong performance across general tasks while maintaining reasonable resource requirements. Good alternative to similar-sized models with competitive capabilities.',
|
|
2297
|
+
},
|
|
2298
|
+
{
|
|
2299
|
+
modelVariant: 'CHAT',
|
|
2300
|
+
modelTitle: 'dolphin-mixtral',
|
|
2301
|
+
modelName: 'dolphin-mixtral',
|
|
2302
|
+
modelDescription: 'Dolphin Mixtral with 32K context window. Community-tuned version of Mixtral with enhanced instruction following and creative capabilities. Maintains the MoE architecture while improving conversational abilities and reducing hallucinations.',
|
|
2303
|
+
},
|
|
2304
|
+
{
|
|
2305
|
+
modelVariant: 'CHAT',
|
|
2306
|
+
modelTitle: 'yi:34b-chat',
|
|
2307
|
+
modelName: 'yi:34b-chat',
|
|
2308
|
+
modelDescription: 'Yi 34B Chat with 4K context window. Large bilingual model with exceptional English and Chinese capabilities. Strong performance on reasoning, knowledge, and instruction following tasks that competes with much larger commercial models.',
|
|
2184
2309
|
},
|
|
2185
2310
|
// <- [🕕]
|
|
2186
2311
|
],
|
|
@@ -2208,7 +2333,7 @@ class OllamaExecutionTools extends OpenAiCompatibleExecutionTools {
|
|
|
2208
2333
|
const openAiCompatibleOptions = {
|
|
2209
2334
|
baseURL: DEFAULT_OLLAMA_BASE_URL,
|
|
2210
2335
|
...ollamaOptions,
|
|
2211
|
-
|
|
2336
|
+
apiKey: 'ollama',
|
|
2212
2337
|
};
|
|
2213
2338
|
super(openAiCompatibleOptions);
|
|
2214
2339
|
}
|
|
@@ -2267,6 +2392,9 @@ class OllamaExecutionTools extends OpenAiCompatibleExecutionTools {
|
|
|
2267
2392
|
* @public exported from `@promptbook/ollama`
|
|
2268
2393
|
*/
|
|
2269
2394
|
const createOllamaExecutionTools = Object.assign((options) => {
|
|
2395
|
+
if (($isRunningInBrowser() || $isRunningInWebWorker()) && !options.dangerouslyAllowBrowser) {
|
|
2396
|
+
options = { ...options, dangerouslyAllowBrowser: true };
|
|
2397
|
+
}
|
|
2270
2398
|
return new OllamaExecutionTools(options);
|
|
2271
2399
|
}, {
|
|
2272
2400
|
packageName: '@promptbook/ollama',
|
|
@@ -2442,7 +2570,7 @@ const $llmToolsRegister = new $Register('llm_execution_tools_constructors');
|
|
|
2442
2570
|
* Warning: This is not useful for the end user, it is just a side effect of the mechanism that handles all available LLM tools
|
|
2443
2571
|
*
|
|
2444
2572
|
* @public exported from `@promptbook/ollama`
|
|
2445
|
-
* @public exported from `@promptbook/
|
|
2573
|
+
* @public exported from `@promptbook/wizard`
|
|
2446
2574
|
* @public exported from `@promptbook/cli`
|
|
2447
2575
|
*/
|
|
2448
2576
|
const _OllamaRegistration = $llmToolsRegister.register(createOllamaExecutionTools);
|