@promptbook/azure-openai 0.94.0-1 → 0.94.0-13

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (27) hide show
  1. package/README.md +6 -8
  2. package/esm/index.es.js +93 -98
  3. package/esm/index.es.js.map +1 -1
  4. package/esm/typings/src/_packages/ollama.index.d.ts +6 -0
  5. package/esm/typings/src/_packages/openai.index.d.ts +2 -0
  6. package/esm/typings/src/execution/AvailableModel.d.ts +9 -1
  7. package/esm/typings/src/llm-providers/_common/filterModels.d.ts +2 -2
  8. package/esm/typings/src/llm-providers/{openai/computeUsage.d.ts → _common/utils/pricing.d.ts} +2 -2
  9. package/esm/typings/src/llm-providers/anthropic-claude/AnthropicClaudeExecutionToolsOptions.d.ts +1 -1
  10. package/esm/typings/src/llm-providers/azure-openai/AzureOpenAiExecutionToolsOptions.d.ts +1 -1
  11. package/esm/typings/src/llm-providers/deepseek/DeepseekExecutionToolsOptions.d.ts +1 -1
  12. package/esm/typings/src/llm-providers/google/GoogleExecutionToolsOptions.d.ts +1 -1
  13. package/esm/typings/src/llm-providers/ollama/OllamaExecutionTools.d.ts +36 -11
  14. package/esm/typings/src/llm-providers/ollama/OllamaExecutionToolsOptions.d.ts +23 -12
  15. package/esm/typings/src/llm-providers/ollama/createOllamaExecutionTools.d.ts +3 -3
  16. package/esm/typings/src/llm-providers/ollama/ollama-models.d.ts +14 -0
  17. package/esm/typings/src/llm-providers/openai/OpenAiAssistantExecutionToolsOptions.d.ts +1 -1
  18. package/esm/typings/src/llm-providers/openai/OpenAiCompatibleExecutionTools.d.ts +91 -0
  19. package/esm/typings/src/llm-providers/openai/OpenAiExecutionTools.d.ts +12 -53
  20. package/esm/typings/src/llm-providers/openai/OpenAiExecutionToolsOptions.d.ts +1 -1
  21. package/esm/typings/src/llm-providers/openai/createOpenAiExecutionTools.d.ts +2 -0
  22. package/esm/typings/src/llm-providers/openai/openai-models.d.ts +1 -7
  23. package/esm/typings/src/version.d.ts +1 -1
  24. package/package.json +25 -2
  25. package/umd/index.umd.js +93 -98
  26. package/umd/index.umd.js.map +1 -1
  27. /package/esm/typings/src/llm-providers/{openai/computeUsage.test.d.ts → _common/utils/pricing.test.d.ts} +0 -0
package/README.md CHANGED
@@ -261,6 +261,8 @@ Rest of the documentation is common for **entire promptbook ecosystem**:
261
261
 
262
262
  During the computer revolution, we have seen [multiple generations of computer languages](https://github.com/webgptorg/promptbook/discussions/180), from the physical rewiring of the vacuum tubes through low-level machine code to the high-level languages like Python or JavaScript. And now, we're on the edge of the **next revolution**!
263
263
 
264
+
265
+
264
266
  It's a revolution of writing software in **plain human language** that is understandable and executable by both humans and machines – and it's going to change everything!
265
267
 
266
268
  The incredible growth in power of microprocessors and the Moore's Law have been the driving force behind the ever-more powerful languages, and it's been an amazing journey! Similarly, the large language models (like GPT or Claude) are the next big thing in language technology, and they're set to transform the way we interact with computers.
@@ -386,16 +388,8 @@ Join our growing community of developers and users:
386
388
 
387
389
  _A concise, Markdown-based DSL for crafting AI workflows and automations._
388
390
 
389
- ---
390
391
 
391
- ### 📑 Table of Contents
392
392
 
393
- - [Introduction](#introduction)
394
- - [Example](#example)
395
- - [1. What: Workflows, Tasks & Parameters](#1-what-workflows-tasks--parameters)
396
- - [2. Who: Personas](#2-who-personas)
397
- - [3. How: Knowledge, Instruments & Actions](#3-how-knowledge-instruments-and-actions)
398
- - [General Principles](#general-principles)
399
393
 
400
394
  ### Introduction
401
395
 
@@ -446,6 +440,8 @@ Personas can have access to different knowledge, tools and actions. They can als
446
440
 
447
441
  - [PERSONA](https://github.com/webgptorg/promptbook/blob/main/documents/commands/PERSONA.md)
448
442
 
443
+
444
+
449
445
  ### **3. How:** Knowledge, Instruments and Actions
450
446
 
451
447
  The resources used by the personas are used to do the work.
@@ -545,6 +541,8 @@ The following glossary is used to clarify certain concepts:
545
541
 
546
542
  _Note: This section is not complete dictionary, more list of general AI / LLM terms that has connection with Promptbook_
547
543
 
544
+
545
+
548
546
  ### 💯 Core concepts
549
547
 
550
548
  - [📚 Collection of pipelines](https://github.com/webgptorg/promptbook/discussions/65)
package/esm/index.es.js CHANGED
@@ -18,7 +18,7 @@ const BOOK_LANGUAGE_VERSION = '1.0.0';
18
18
  * @generated
19
19
  * @see https://github.com/webgptorg/promptbook
20
20
  */
21
- const PROMPTBOOK_ENGINE_VERSION = '0.94.0-1';
21
+ const PROMPTBOOK_ENGINE_VERSION = '0.94.0-13';
22
22
  /**
23
23
  * TODO: string_promptbook_version should be constrained to the all versions of Promptbook engine
24
24
  * Note: [💞] Ignore a discrepancy between file name and entity name
@@ -1194,11 +1194,11 @@ function templateParameters(template, parameters) {
1194
1194
  }
1195
1195
 
1196
1196
  /**
1197
- * Function computeUsage will create price per one token based on the string value found on openai page
1197
+ * Create price per one token based on the string value found on openai page
1198
1198
  *
1199
1199
  * @private within the repository, used only as internal helper for `OPENAI_MODELS`
1200
1200
  */
1201
- function computeUsage(value) {
1201
+ function pricing(value) {
1202
1202
  const [price, tokens] = value.split(' / ');
1203
1203
  return parseFloat(price.replace('$', '')) / parseFloat(tokens.replace('M tokens', '')) / 1000000;
1204
1204
  }
@@ -1232,10 +1232,10 @@ const OPENAI_MODELS = exportJson({
1232
1232
  modelVariant: 'COMPLETION',
1233
1233
  modelTitle: 'davinci-002',
1234
1234
  modelName: 'davinci-002',
1235
- modelDescription: 'Legacy completion model with strong performance on text generation tasks. Optimized for complex instructions and longer outputs.',
1235
+ modelDescription: 'Legacy completion model with 4K token context window. Excels at complex text generation, creative writing, and detailed content creation with strong contextual understanding. Optimized for instructions requiring nuanced outputs and extended reasoning. Suitable for applications needing high-quality text generation without conversation management.',
1236
1236
  pricing: {
1237
- prompt: computeUsage(`$2.00 / 1M tokens`),
1238
- output: computeUsage(`$2.00 / 1M tokens`),
1237
+ prompt: pricing(`$2.00 / 1M tokens`),
1238
+ output: pricing(`$2.00 / 1M tokens`),
1239
1239
  },
1240
1240
  },
1241
1241
  /**/
@@ -1250,10 +1250,10 @@ const OPENAI_MODELS = exportJson({
1250
1250
  modelVariant: 'CHAT',
1251
1251
  modelTitle: 'gpt-3.5-turbo-16k',
1252
1252
  modelName: 'gpt-3.5-turbo-16k',
1253
- modelDescription: 'GPT-3.5 Turbo with extended 16k token context length for handling longer conversations and documents.',
1253
+ modelDescription: 'Extended context GPT-3.5 Turbo with 16K token window. Maintains core capabilities of standard 3.5 Turbo while supporting longer conversations and documents. Features good balance of performance and cost for applications requiring more context than standard 4K models. Effective for document analysis, extended conversations, and multi-step reasoning tasks.',
1254
1254
  pricing: {
1255
- prompt: computeUsage(`$3.00 / 1M tokens`),
1256
- output: computeUsage(`$4.00 / 1M tokens`),
1255
+ prompt: pricing(`$3.00 / 1M tokens`),
1256
+ output: pricing(`$4.00 / 1M tokens`),
1257
1257
  },
1258
1258
  },
1259
1259
  /**/
@@ -1274,10 +1274,10 @@ const OPENAI_MODELS = exportJson({
1274
1274
  modelVariant: 'CHAT',
1275
1275
  modelTitle: 'gpt-4',
1276
1276
  modelName: 'gpt-4',
1277
- modelDescription: 'GPT-4 is a powerful language model with enhanced reasoning, instruction-following capabilities, and 8K context window. Optimized for complex tasks requiring deep understanding.',
1277
+ modelDescription: 'Powerful language model with 8K context window featuring sophisticated reasoning, instruction-following, and knowledge capabilities. Demonstrates strong performance on complex tasks requiring deep understanding and multi-step reasoning. Excels at code generation, logical analysis, and nuanced content creation. Suitable for advanced applications requiring high-quality outputs.',
1278
1278
  pricing: {
1279
- prompt: computeUsage(`$30.00 / 1M tokens`),
1280
- output: computeUsage(`$60.00 / 1M tokens`),
1279
+ prompt: pricing(`$30.00 / 1M tokens`),
1280
+ output: pricing(`$60.00 / 1M tokens`),
1281
1281
  },
1282
1282
  },
1283
1283
  /**/
@@ -1286,10 +1286,10 @@ const OPENAI_MODELS = exportJson({
1286
1286
  modelVariant: 'CHAT',
1287
1287
  modelTitle: 'gpt-4-32k',
1288
1288
  modelName: 'gpt-4-32k',
1289
- modelDescription: 'Extended context version of GPT-4 with a 32K token window for processing very long inputs and generating comprehensive responses for complex tasks.',
1289
+ modelDescription: 'Extended context version of GPT-4 with 32K token window. Maintains all capabilities of standard GPT-4 while supporting analysis of very lengthy documents, code bases, and conversations. Features enhanced ability to maintain context over long interactions and process detailed information from large inputs. Ideal for document analysis, legal review, and complex problem-solving.',
1290
1290
  pricing: {
1291
- prompt: computeUsage(`$60.00 / 1M tokens`),
1292
- output: computeUsage(`$120.00 / 1M tokens`),
1291
+ prompt: pricing(`$60.00 / 1M tokens`),
1292
+ output: pricing(`$120.00 / 1M tokens`),
1293
1293
  },
1294
1294
  },
1295
1295
  /**/
@@ -1309,10 +1309,10 @@ const OPENAI_MODELS = exportJson({
1309
1309
  modelVariant: 'CHAT',
1310
1310
  modelTitle: 'gpt-4-turbo-2024-04-09',
1311
1311
  modelName: 'gpt-4-turbo-2024-04-09',
1312
- modelDescription: 'Latest stable GPT-4 Turbo model from April 2024 with enhanced reasoning and context handling capabilities. Offers 128K context window and improved performance.',
1312
+ modelDescription: 'Latest stable GPT-4 Turbo from April 2024 with 128K context window. Features enhanced reasoning chains, improved factual accuracy with 40% reduction in hallucinations, and better instruction following compared to earlier versions. Includes advanced function calling capabilities and knowledge up to April 2024. Provides optimal performance for enterprise applications requiring reliability.',
1313
1313
  pricing: {
1314
- prompt: computeUsage(`$10.00 / 1M tokens`),
1315
- output: computeUsage(`$30.00 / 1M tokens`),
1314
+ prompt: pricing(`$10.00 / 1M tokens`),
1315
+ output: pricing(`$30.00 / 1M tokens`),
1316
1316
  },
1317
1317
  },
1318
1318
  /**/
@@ -1321,10 +1321,10 @@ const OPENAI_MODELS = exportJson({
1321
1321
  modelVariant: 'CHAT',
1322
1322
  modelTitle: 'gpt-3.5-turbo-1106',
1323
1323
  modelName: 'gpt-3.5-turbo-1106',
1324
- modelDescription: 'November 2023 version of GPT-3.5 Turbo with improved instruction following and a 16K token context window.',
1324
+ modelDescription: 'November 2023 version of GPT-3.5 Turbo with 16K token context window. Features improved instruction following, more consistent output formatting, and enhanced function calling capabilities. Includes knowledge cutoff from April 2023. Suitable for applications requiring good performance at lower cost than GPT-4 models.',
1325
1325
  pricing: {
1326
- prompt: computeUsage(`$1.00 / 1M tokens`),
1327
- output: computeUsage(`$2.00 / 1M tokens`),
1326
+ prompt: pricing(`$1.00 / 1M tokens`),
1327
+ output: pricing(`$2.00 / 1M tokens`),
1328
1328
  },
1329
1329
  },
1330
1330
  /**/
@@ -1333,10 +1333,10 @@ const OPENAI_MODELS = exportJson({
1333
1333
  modelVariant: 'CHAT',
1334
1334
  modelTitle: 'gpt-4-turbo',
1335
1335
  modelName: 'gpt-4-turbo',
1336
- modelDescription: 'More capable model than GPT-4 with improved instruction following, function calling and a 128K token context window for handling very large documents.',
1336
+ modelDescription: 'More capable and cost-efficient version of GPT-4 with 128K token context window. Features improved instruction following, advanced function calling capabilities, and better performance on coding tasks. Maintains superior reasoning and knowledge while offering substantial cost reduction compared to base GPT-4. Ideal for complex applications requiring extensive context processing.',
1337
1337
  pricing: {
1338
- prompt: computeUsage(`$10.00 / 1M tokens`),
1339
- output: computeUsage(`$30.00 / 1M tokens`),
1338
+ prompt: pricing(`$10.00 / 1M tokens`),
1339
+ output: pricing(`$30.00 / 1M tokens`),
1340
1340
  },
1341
1341
  },
1342
1342
  /**/
@@ -1345,10 +1345,10 @@ const OPENAI_MODELS = exportJson({
1345
1345
  modelVariant: 'COMPLETION',
1346
1346
  modelTitle: 'gpt-3.5-turbo-instruct-0914',
1347
1347
  modelName: 'gpt-3.5-turbo-instruct-0914',
1348
- modelDescription: 'September 2023 version of GPT-3.5 Turbo optimized for completion-style instruction following with a 4K context window.',
1348
+ modelDescription: 'September 2023 version of GPT-3.5 Turbo Instruct with 4K context window. Optimized for completion-style instruction following with deterministic responses. Better suited than chat models for applications requiring specific formatted outputs without conversation management. Knowledge cutoff from September 2021.',
1349
1349
  pricing: {
1350
- prompt: computeUsage(`$1.50 / 1M tokens`),
1351
- output: computeUsage(`$2.00 / 1M tokens`), // <- For gpt-3.5-turbo-instruct
1350
+ prompt: pricing(`$1.50 / 1M tokens`),
1351
+ output: pricing(`$2.00 / 1M tokens`),
1352
1352
  },
1353
1353
  },
1354
1354
  /**/
@@ -1357,10 +1357,10 @@ const OPENAI_MODELS = exportJson({
1357
1357
  modelVariant: 'COMPLETION',
1358
1358
  modelTitle: 'gpt-3.5-turbo-instruct',
1359
1359
  modelName: 'gpt-3.5-turbo-instruct',
1360
- modelDescription: 'Optimized version of GPT-3.5 for completion-style API with good instruction following and a 4K token context window.',
1360
+ modelDescription: 'Optimized version of GPT-3.5 for completion-style API with 4K token context window. Features strong instruction following with single-turn design rather than multi-turn conversation. Provides more consistent, deterministic outputs compared to chat models. Well-suited for templated content generation and structured text transformation tasks.',
1361
1361
  pricing: {
1362
- prompt: computeUsage(`$1.50 / 1M tokens`),
1363
- output: computeUsage(`$2.00 / 1M tokens`),
1362
+ prompt: pricing(`$1.50 / 1M tokens`),
1363
+ output: pricing(`$2.00 / 1M tokens`),
1364
1364
  },
1365
1365
  },
1366
1366
  /**/
@@ -1375,10 +1375,10 @@ const OPENAI_MODELS = exportJson({
1375
1375
  modelVariant: 'CHAT',
1376
1376
  modelTitle: 'gpt-3.5-turbo',
1377
1377
  modelName: 'gpt-3.5-turbo',
1378
- modelDescription: 'Latest version of GPT-3.5 Turbo with improved performance and instruction following capabilities. Default 4K context window with options for 16K.',
1378
+ modelDescription: 'Latest version of GPT-3.5 Turbo with 4K token default context window (16K available). Features continually improved performance with enhanced instruction following and reduced hallucinations. Offers excellent balance between capability and cost efficiency. Suitable for most general-purpose applications requiring good AI capabilities at reasonable cost.',
1379
1379
  pricing: {
1380
- prompt: computeUsage(`$0.50 / 1M tokens`),
1381
- output: computeUsage(`$1.50 / 1M tokens`),
1380
+ prompt: pricing(`$0.50 / 1M tokens`),
1381
+ output: pricing(`$1.50 / 1M tokens`),
1382
1382
  },
1383
1383
  },
1384
1384
  /**/
@@ -1387,10 +1387,10 @@ const OPENAI_MODELS = exportJson({
1387
1387
  modelVariant: 'CHAT',
1388
1388
  modelTitle: 'gpt-3.5-turbo-0301',
1389
1389
  modelName: 'gpt-3.5-turbo-0301',
1390
- modelDescription: 'March 2023 version of GPT-3.5 Turbo with a 4K token context window. Legacy model maintained for backward compatibility.',
1390
+ modelDescription: 'March 2023 version of GPT-3.5 Turbo with 4K token context window. Legacy model maintained for backward compatibility with specific application behaviors. Features solid conversational abilities and basic instruction following. Knowledge cutoff from September 2021. Suitable for applications explicitly designed for this version.',
1391
1391
  pricing: {
1392
- prompt: computeUsage(`$1.50 / 1M tokens`),
1393
- output: computeUsage(`$2.00 / 1M tokens`),
1392
+ prompt: pricing(`$1.50 / 1M tokens`),
1393
+ output: pricing(`$2.00 / 1M tokens`),
1394
1394
  },
1395
1395
  },
1396
1396
  /**/
@@ -1399,10 +1399,10 @@ const OPENAI_MODELS = exportJson({
1399
1399
  modelVariant: 'COMPLETION',
1400
1400
  modelTitle: 'babbage-002',
1401
1401
  modelName: 'babbage-002',
1402
- modelDescription: 'Efficient legacy completion model with a good balance of performance and speed. Suitable for straightforward text generation tasks.',
1402
+ modelDescription: 'Efficient legacy completion model with 4K context window balancing performance and speed. Features moderate reasoning capabilities with focus on straightforward text generation tasks. Significantly more efficient than davinci models while maintaining adequate quality for many applications. Suitable for high-volume, cost-sensitive text generation needs.',
1403
1403
  pricing: {
1404
- prompt: computeUsage(`$0.40 / 1M tokens`),
1405
- output: computeUsage(`$0.40 / 1M tokens`),
1404
+ prompt: pricing(`$0.40 / 1M tokens`),
1405
+ output: pricing(`$0.40 / 1M tokens`),
1406
1406
  },
1407
1407
  },
1408
1408
  /**/
@@ -1411,10 +1411,10 @@ const OPENAI_MODELS = exportJson({
1411
1411
  modelVariant: 'CHAT',
1412
1412
  modelTitle: 'gpt-4-1106-preview',
1413
1413
  modelName: 'gpt-4-1106-preview',
1414
- modelDescription: 'November 2023 preview version of GPT-4 Turbo with improved instruction following and a 128K token context window.',
1414
+ modelDescription: 'November 2023 preview version of GPT-4 Turbo with 128K token context window. Features improved instruction following, better function calling capabilities, and enhanced reasoning. Includes knowledge cutoff from April 2023. Suitable for complex applications requiring extensive document understanding and sophisticated interactions.',
1415
1415
  pricing: {
1416
- prompt: computeUsage(`$10.00 / 1M tokens`),
1417
- output: computeUsage(`$30.00 / 1M tokens`),
1416
+ prompt: pricing(`$10.00 / 1M tokens`),
1417
+ output: pricing(`$30.00 / 1M tokens`),
1418
1418
  },
1419
1419
  },
1420
1420
  /**/
@@ -1423,10 +1423,10 @@ const OPENAI_MODELS = exportJson({
1423
1423
  modelVariant: 'CHAT',
1424
1424
  modelTitle: 'gpt-4-0125-preview',
1425
1425
  modelName: 'gpt-4-0125-preview',
1426
- modelDescription: 'January 2024 preview version of GPT-4 Turbo with improved reasoning capabilities and a 128K token context window.',
1426
+ modelDescription: 'January 2024 preview version of GPT-4 Turbo with 128K token context window. Features improved reasoning capabilities, enhanced tool use, and more reliable function calling. Includes knowledge cutoff from October 2023. Offers better performance on complex logical tasks and more consistent outputs than previous preview versions.',
1427
1427
  pricing: {
1428
- prompt: computeUsage(`$10.00 / 1M tokens`),
1429
- output: computeUsage(`$30.00 / 1M tokens`),
1428
+ prompt: pricing(`$10.00 / 1M tokens`),
1429
+ output: pricing(`$30.00 / 1M tokens`),
1430
1430
  },
1431
1431
  },
1432
1432
  /**/
@@ -1441,10 +1441,10 @@ const OPENAI_MODELS = exportJson({
1441
1441
  modelVariant: 'CHAT',
1442
1442
  modelTitle: 'gpt-3.5-turbo-0125',
1443
1443
  modelName: 'gpt-3.5-turbo-0125',
1444
- modelDescription: 'January 2024 version of GPT-3.5 Turbo with improved reasoning capabilities and a 16K token context window.',
1444
+ modelDescription: 'January 2024 version of GPT-3.5 Turbo with 16K token context window. Features improved reasoning capabilities, better instruction adherence, and reduced hallucinations compared to previous versions. Includes knowledge cutoff from September 2021. Provides good performance for most general applications at reasonable cost.',
1445
1445
  pricing: {
1446
- prompt: computeUsage(`$0.50 / 1M tokens`),
1447
- output: computeUsage(`$1.50 / 1M tokens`),
1446
+ prompt: pricing(`$0.50 / 1M tokens`),
1447
+ output: pricing(`$1.50 / 1M tokens`),
1448
1448
  },
1449
1449
  },
1450
1450
  /**/
@@ -1453,10 +1453,10 @@ const OPENAI_MODELS = exportJson({
1453
1453
  modelVariant: 'CHAT',
1454
1454
  modelTitle: 'gpt-4-turbo-preview',
1455
1455
  modelName: 'gpt-4-turbo-preview',
1456
- modelDescription: 'Preview version of GPT-4 Turbo that points to the latest model version. Features improved instruction following, 128K token context window and lower latency.',
1456
+ modelDescription: 'Preview version of GPT-4 Turbo with 128K token context window that points to the latest development model. Features cutting-edge improvements to instruction following, knowledge representation, and tool use capabilities. Provides access to newest features but may have occasional behavior changes. Best for non-critical applications wanting latest capabilities.',
1457
1457
  pricing: {
1458
- prompt: computeUsage(`$10.00 / 1M tokens`),
1459
- output: computeUsage(`$30.00 / 1M tokens`),
1458
+ prompt: pricing(`$10.00 / 1M tokens`),
1459
+ output: pricing(`$30.00 / 1M tokens`),
1460
1460
  },
1461
1461
  },
1462
1462
  /**/
@@ -1465,11 +1465,10 @@ const OPENAI_MODELS = exportJson({
1465
1465
  modelVariant: 'EMBEDDING',
1466
1466
  modelTitle: 'text-embedding-3-large',
1467
1467
  modelName: 'text-embedding-3-large',
1468
- modelDescription: "OpenAI's most capable text embedding model designed for high-quality embeddings for complex similarity tasks and information retrieval.",
1468
+ modelDescription: "OpenAI's most capable text embedding model generating 3072-dimensional vectors. Designed for high-quality embeddings for complex similarity tasks, clustering, and information retrieval. Features enhanced cross-lingual capabilities and significantly improved performance on retrieval and classification benchmarks. Ideal for sophisticated RAG systems and semantic search applications.",
1469
1469
  pricing: {
1470
- prompt: computeUsage(`$0.13 / 1M tokens`),
1471
- // TODO: [🏏] Leverage the batch API @see https://platform.openai.com/docs/guides/batch
1472
- output: 0, // <- Note: [🆖] In Embedding models you dont pay for output
1470
+ prompt: pricing(`$0.13 / 1M tokens`),
1471
+ output: 0,
1473
1472
  },
1474
1473
  },
1475
1474
  /**/
@@ -1478,11 +1477,10 @@ const OPENAI_MODELS = exportJson({
1478
1477
  modelVariant: 'EMBEDDING',
1479
1478
  modelTitle: 'text-embedding-3-small',
1480
1479
  modelName: 'text-embedding-3-small',
1481
- modelDescription: 'Cost-effective embedding model with good performance for simpler tasks like text similarity and retrieval. Good balance of quality and efficiency.',
1480
+ modelDescription: 'Cost-effective embedding model generating 1536-dimensional vectors. Balances quality and efficiency for simpler tasks while maintaining good performance on text similarity and retrieval applications. Offers 20% better quality than ada-002 at significantly lower cost. Ideal for production embedding applications with cost constraints.',
1482
1481
  pricing: {
1483
- prompt: computeUsage(`$0.02 / 1M tokens`),
1484
- // TODO: [🏏] Leverage the batch API @see https://platform.openai.com/docs/guides/batch
1485
- output: 0, // <- Note: [🆖] In Embedding models you dont pay for output
1482
+ prompt: pricing(`$0.02 / 1M tokens`),
1483
+ output: 0,
1486
1484
  },
1487
1485
  },
1488
1486
  /**/
@@ -1491,10 +1489,10 @@ const OPENAI_MODELS = exportJson({
1491
1489
  modelVariant: 'CHAT',
1492
1490
  modelTitle: 'gpt-3.5-turbo-0613',
1493
1491
  modelName: 'gpt-3.5-turbo-0613',
1494
- modelDescription: 'June 2023 version of GPT-3.5 Turbo with function calling capabilities and a 4K token context window.',
1492
+ modelDescription: "June 2023 version of GPT-3.5 Turbo with 4K token context window. Features function calling capabilities for structured data extraction and API interaction. Includes knowledge cutoff from September 2021. Maintained for applications specifically designed for this version's behaviors and capabilities.",
1495
1493
  pricing: {
1496
- prompt: computeUsage(`$1.50 / 1M tokens`),
1497
- output: computeUsage(`$2.00 / 1M tokens`),
1494
+ prompt: pricing(`$1.50 / 1M tokens`),
1495
+ output: pricing(`$2.00 / 1M tokens`),
1498
1496
  },
1499
1497
  },
1500
1498
  /**/
@@ -1503,11 +1501,10 @@ const OPENAI_MODELS = exportJson({
1503
1501
  modelVariant: 'EMBEDDING',
1504
1502
  modelTitle: 'text-embedding-ada-002',
1505
1503
  modelName: 'text-embedding-ada-002',
1506
- modelDescription: 'Legacy text embedding model suitable for text similarity and retrieval augmented generation use cases. Replaced by newer embedding-3 models.',
1504
+ modelDescription: 'Legacy text embedding model generating 1536-dimensional vectors suitable for text similarity and retrieval applications. Processes up to 8K tokens per request with consistent embedding quality. While superseded by newer embedding-3 models, still maintains adequate performance for many semantic search and classification tasks.',
1507
1505
  pricing: {
1508
- prompt: computeUsage(`$0.1 / 1M tokens`),
1509
- // TODO: [🏏] Leverage the batch API @see https://platform.openai.com/docs/guides/batch
1510
- output: 0, // <- Note: [🆖] In Embedding models you dont pay for output
1506
+ prompt: pricing(`$0.1 / 1M tokens`),
1507
+ output: 0,
1511
1508
  },
1512
1509
  },
1513
1510
  /**/
@@ -1534,10 +1531,10 @@ const OPENAI_MODELS = exportJson({
1534
1531
  modelVariant: 'CHAT',
1535
1532
  modelTitle: 'gpt-4o-2024-05-13',
1536
1533
  modelName: 'gpt-4o-2024-05-13',
1537
- modelDescription: 'May 2024 version of GPT-4o with enhanced multimodal capabilities, improved reasoning, and optimized for vision, audio and chat at lower latencies.',
1534
+ modelDescription: 'May 2024 version of GPT-4o with 128K context window. Features enhanced multimodal capabilities including superior image understanding (up to 20MP), audio processing, and improved reasoning. Optimized for 2x lower latency than GPT-4 Turbo while maintaining high performance. Includes knowledge up to October 2023. Ideal for production applications requiring reliable multimodal capabilities.',
1538
1535
  pricing: {
1539
- prompt: computeUsage(`$5.00 / 1M tokens`),
1540
- output: computeUsage(`$15.00 / 1M tokens`),
1536
+ prompt: pricing(`$5.00 / 1M tokens`),
1537
+ output: pricing(`$15.00 / 1M tokens`),
1541
1538
  },
1542
1539
  },
1543
1540
  /**/
@@ -1546,10 +1543,10 @@ const OPENAI_MODELS = exportJson({
1546
1543
  modelVariant: 'CHAT',
1547
1544
  modelTitle: 'gpt-4o',
1548
1545
  modelName: 'gpt-4o',
1549
- modelDescription: "OpenAI's most advanced multimodal model optimized for performance, speed, and cost. Capable of vision, reasoning, and high quality text generation.",
1546
+ modelDescription: "OpenAI's most advanced general-purpose multimodal model with 128K context window. Optimized for balanced performance, speed, and cost with 2x faster responses than GPT-4 Turbo. Features excellent vision processing, audio understanding, reasoning, and text generation quality. Represents optimal balance of capability and efficiency for most advanced applications.",
1550
1547
  pricing: {
1551
- prompt: computeUsage(`$5.00 / 1M tokens`),
1552
- output: computeUsage(`$15.00 / 1M tokens`),
1548
+ prompt: pricing(`$5.00 / 1M tokens`),
1549
+ output: pricing(`$15.00 / 1M tokens`),
1553
1550
  },
1554
1551
  },
1555
1552
  /**/
@@ -1558,10 +1555,10 @@ const OPENAI_MODELS = exportJson({
1558
1555
  modelVariant: 'CHAT',
1559
1556
  modelTitle: 'gpt-4o-mini',
1560
1557
  modelName: 'gpt-4o-mini',
1561
- modelDescription: 'Smaller, more cost-effective version of GPT-4o with good performance across text, vision, and audio tasks at reduced complexity.',
1558
+ modelDescription: 'Smaller, more cost-effective version of GPT-4o with 128K context window. Maintains impressive capabilities across text, vision, and audio tasks while operating at significantly lower cost. Features 3x faster inference than GPT-4o with good performance on general tasks. Excellent for applications requiring good quality multimodal capabilities at scale.',
1562
1559
  pricing: {
1563
- prompt: computeUsage(`$0.15 / 1M tokens`),
1564
- output: computeUsage(`$0.60 / 1M tokens`),
1560
+ prompt: pricing(`$0.15 / 1M tokens`),
1561
+ output: pricing(`$0.60 / 1M tokens`),
1565
1562
  },
1566
1563
  },
1567
1564
  /**/
@@ -1570,10 +1567,10 @@ const OPENAI_MODELS = exportJson({
1570
1567
  modelVariant: 'CHAT',
1571
1568
  modelTitle: 'o1-preview',
1572
1569
  modelName: 'o1-preview',
1573
- modelDescription: 'Advanced reasoning model with exceptional performance on complex logical, mathematical, and analytical tasks. Built for deep reasoning and specialized professional tasks.',
1570
+ modelDescription: 'Advanced reasoning model with 128K context window specializing in complex logical, mathematical, and analytical tasks. Features exceptional step-by-step problem-solving capabilities, advanced mathematical and scientific reasoning, and superior performance on STEM-focused problems. Significantly outperforms GPT-4 on quantitative reasoning benchmarks. Ideal for professional and specialized applications.',
1574
1571
  pricing: {
1575
- prompt: computeUsage(`$15.00 / 1M tokens`),
1576
- output: computeUsage(`$60.00 / 1M tokens`),
1572
+ prompt: pricing(`$15.00 / 1M tokens`),
1573
+ output: pricing(`$60.00 / 1M tokens`),
1577
1574
  },
1578
1575
  },
1579
1576
  /**/
@@ -1582,11 +1579,10 @@ const OPENAI_MODELS = exportJson({
1582
1579
  modelVariant: 'CHAT',
1583
1580
  modelTitle: 'o1-preview-2024-09-12',
1584
1581
  modelName: 'o1-preview-2024-09-12',
1585
- modelDescription: 'September 2024 version of O1 preview with specialized reasoning capabilities for complex tasks requiring precise analytical thinking.',
1586
- // <- TODO: [💩] Some better system to organize these date suffixes and versions
1582
+ modelDescription: 'September 2024 version of O1 preview with 128K context window. Features specialized reasoning capabilities with 30% improvement on mathematical and scientific accuracy over previous versions. Includes enhanced support for formal logic, statistical analysis, and technical domains. Optimized for professional applications requiring precise analytical thinking and rigorous methodologies.',
1587
1583
  pricing: {
1588
- prompt: computeUsage(`$15.00 / 1M tokens`),
1589
- output: computeUsage(`$60.00 / 1M tokens`),
1584
+ prompt: pricing(`$15.00 / 1M tokens`),
1585
+ output: pricing(`$60.00 / 1M tokens`),
1590
1586
  },
1591
1587
  },
1592
1588
  /**/
@@ -1595,10 +1591,10 @@ const OPENAI_MODELS = exportJson({
1595
1591
  modelVariant: 'CHAT',
1596
1592
  modelTitle: 'o1-mini',
1597
1593
  modelName: 'o1-mini',
1598
- modelDescription: 'Smaller, cost-effective version of the O1 model with good performance on reasoning tasks while maintaining efficiency for everyday analytical use.',
1594
+ modelDescription: 'Smaller, cost-effective version of the O1 model with 128K context window. Maintains strong analytical reasoning abilities while reducing computational requirements by 70%. Features good performance on mathematical, logical, and scientific tasks at significantly lower cost than full O1. Excellent for everyday analytical applications that benefit from reasoning focus.',
1599
1595
  pricing: {
1600
- prompt: computeUsage(`$3.00 / 1M tokens`),
1601
- output: computeUsage(`$12.00 / 1M tokens`),
1596
+ prompt: pricing(`$3.00 / 1M tokens`),
1597
+ output: pricing(`$12.00 / 1M tokens`),
1602
1598
  },
1603
1599
  },
1604
1600
  /**/
@@ -1607,10 +1603,10 @@ const OPENAI_MODELS = exportJson({
1607
1603
  modelVariant: 'CHAT',
1608
1604
  modelTitle: 'o1',
1609
1605
  modelName: 'o1',
1610
- modelDescription: "OpenAI's advanced reasoning model focused on logic and problem-solving. Designed for complex analytical tasks with rigorous step-by-step reasoning. 128K context window.",
1606
+ modelDescription: "OpenAI's advanced reasoning model with 128K context window focusing on logical problem-solving and analytical thinking. Features exceptional performance on quantitative tasks, step-by-step deduction, and complex technical problems. Maintains 95%+ of o1-preview capabilities with production-ready stability. Ideal for scientific computing, financial analysis, and professional applications.",
1611
1607
  pricing: {
1612
- prompt: computeUsage(`$15.00 / 1M tokens`),
1613
- output: computeUsage(`$60.00 / 1M tokens`),
1608
+ prompt: pricing(`$15.00 / 1M tokens`),
1609
+ output: pricing(`$60.00 / 1M tokens`),
1614
1610
  },
1615
1611
  },
1616
1612
  /**/
@@ -1619,11 +1615,10 @@ const OPENAI_MODELS = exportJson({
1619
1615
  modelVariant: 'CHAT',
1620
1616
  modelTitle: 'o3-mini',
1621
1617
  modelName: 'o3-mini',
1622
- modelDescription: 'Cost-effective reasoning model optimized for academic and scientific problem-solving. Efficient performance on STEM tasks with deep mathematical and scientific knowledge. 128K context window.',
1618
+ modelDescription: 'Cost-effective reasoning model with 128K context window optimized for academic and scientific problem-solving. Features efficient performance on STEM tasks with specialized capabilities in mathematics, physics, chemistry, and computer science. Offers 80% of O1 performance on technical domains at significantly lower cost. Ideal for educational applications and research support.',
1623
1619
  pricing: {
1624
- prompt: computeUsage(`$3.00 / 1M tokens`),
1625
- output: computeUsage(`$12.00 / 1M tokens`),
1626
- // <- TODO: !! Unsure, check the pricing
1620
+ prompt: pricing(`$3.00 / 1M tokens`),
1621
+ output: pricing(`$12.00 / 1M tokens`),
1627
1622
  },
1628
1623
  },
1629
1624
  /**/
@@ -1632,10 +1627,10 @@ const OPENAI_MODELS = exportJson({
1632
1627
  modelVariant: 'CHAT',
1633
1628
  modelTitle: 'o1-mini-2024-09-12',
1634
1629
  modelName: 'o1-mini-2024-09-12',
1635
- modelDescription: "September 2024 version of O1-mini with balanced reasoning capabilities and cost-efficiency. Good for analytical tasks that don't require the full O1 model.",
1630
+ modelDescription: "September 2024 version of O1-mini with 128K context window featuring balanced reasoning capabilities and cost-efficiency. Includes 25% improvement in mathematical accuracy and enhanced performance on coding tasks compared to previous versions. Maintains efficient resource utilization while delivering improved results for analytical applications that don't require the full O1 model.",
1636
1631
  pricing: {
1637
- prompt: computeUsage(`$3.00 / 1M tokens`),
1638
- output: computeUsage(`$12.00 / 1M tokens`),
1632
+ prompt: pricing(`$3.00 / 1M tokens`),
1633
+ output: pricing(`$12.00 / 1M tokens`),
1639
1634
  },
1640
1635
  },
1641
1636
  /**/
@@ -1644,10 +1639,10 @@ const OPENAI_MODELS = exportJson({
1644
1639
  modelVariant: 'CHAT',
1645
1640
  modelTitle: 'gpt-3.5-turbo-16k-0613',
1646
1641
  modelName: 'gpt-3.5-turbo-16k-0613',
1647
- modelDescription: 'June 2023 version of GPT-3.5 Turbo with extended 16k token context window for processing longer conversations and documents.',
1642
+ modelDescription: "June 2023 version of GPT-3.5 Turbo with extended 16K token context window. Features good handling of longer conversations and documents with improved memory management across extended contexts. Includes knowledge cutoff from September 2021. Maintained for applications specifically designed for this version's behaviors and capabilities.",
1648
1643
  pricing: {
1649
- prompt: computeUsage(`$3.00 / 1M tokens`),
1650
- output: computeUsage(`$4.00 / 1M tokens`),
1644
+ prompt: pricing(`$3.00 / 1M tokens`),
1645
+ output: pricing(`$4.00 / 1M tokens`),
1651
1646
  },
1652
1647
  },
1653
1648
  /**/