@botpress/cognitive 0.1.42 → 0.1.44

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.d.ts CHANGED
@@ -18979,7 +18979,7 @@ declare class Cognitive {
18979
18979
  private _generateContent;
18980
18980
  }
18981
18981
 
18982
- type Models = 'auto' | 'best' | 'fast' | 'reasoning' | 'cheapest' | 'balance' | 'recommended' | 'reasoning' | 'general-purpose' | 'low-cost' | 'vision' | 'coding' | 'function-calling' | 'agents' | 'storytelling' | 'preview' | 'roleplay' | 'anthropic:claude-3-5-haiku-20241022' | 'anthropic:claude-3-5-sonnet-20240620' | 'anthropic:claude-3-5-sonnet-20241022' | 'anthropic:claude-3-7-sonnet-20250219' | 'anthropic:claude-3-haiku-20240307' | 'anthropic:claude-sonnet-4-20250514' | 'cerebras:gpt-oss-120b' | 'cerebras:llama-4-scout-17b-16e-instruct' | 'cerebras:llama3.1-8b' | 'cerebras:llama3.3-70b' | 'cerebras:qwen-3-32b' | 'fireworks:deepseek-r1' | 'fireworks:deepseek-r1-0528' | 'fireworks:deepseek-r1-basic' | 'fireworks:deepseek-v3-0324' | 'fireworks:gpt-oss-120b' | 'fireworks:gpt-oss-20b' | 'fireworks:llama-v3p1-8b-instruct' | 'fireworks:llama-v3p3-70b-instruct' | 'fireworks:llama4-maverick-instruct-basic' | 'fireworks:llama4-scout-instruct-basic' | 'fireworks:mixtral-8x22b-instruct' | 'fireworks:mixtral-8x7b-instruct' | 'fireworks:mythomax-l2-13b' | 'google-ai:gemini-2.5-flash' | 'google-ai:gemini-2.5-pro' | 'google-ai:models/gemini-2.0-flash' | 'groq:deepseek-r1-distill-llama-70b' | 'groq:gemma2-9b-it' | 'groq:llama-3.1-8b-instant' | 'groq:llama-3.3-70b-versatile' | 'groq:openai/gpt-oss-120b' | 'groq:openai/gpt-oss-20b' | 'openai:gpt-4.1-2025-04-14' | 'openai:gpt-4.1-mini-2025-04-14' | 'openai:gpt-4.1-nano-2025-04-14' | 'openai:gpt-4o-2024-11-20' | 'openai:gpt-4o-mini-2024-07-18' | 'openai:gpt-5-2025-08-07' | 'openai:gpt-5-mini-2025-08-07' | 'openai:gpt-5-nano-2025-08-07' | 'openai:o1-2024-12-17' | 'openai:o1-mini-2024-09-12' | 'openai:o3-2025-04-16' | 'openai:o3-mini-2025-01-31' | 'openai:o4-mini-2025-04-16' | 'openrouter:gpt-oss-120b' | ({} & string);
18982
+ type Models = 'auto' | 'best' | 'fast' | 'reasoning' | 'cheapest' | 'balance' | 'recommended' | 'reasoning' | 'general-purpose' | 'low-cost' | 'vision' | 'coding' | 'function-calling' | 'agents' | 'storytelling' | 'preview' | 'roleplay' | 'anthropic:claude-3-5-haiku-20241022' | 'anthropic:claude-3-5-sonnet-20240620' | 'anthropic:claude-3-5-sonnet-20241022' | 'anthropic:claude-3-7-sonnet-20250219' | 'anthropic:claude-3-haiku-20240307' | 'anthropic:claude-sonnet-4-20250514' | 'cerebras:gpt-oss-120b' | 'cerebras:llama-4-scout-17b-16e-instruct' | 'cerebras:llama3.1-8b' | 'cerebras:llama3.3-70b' | 'cerebras:qwen-3-32b' | 'fireworks-ai:deepseek-r1' | 'fireworks-ai:deepseek-r1-0528' | 'fireworks-ai:deepseek-r1-basic' | 'fireworks-ai:deepseek-v3-0324' | 'fireworks-ai:gpt-oss-120b' | 'fireworks-ai:gpt-oss-20b' | 'fireworks-ai:llama-v3p1-8b-instruct' | 'fireworks-ai:llama-v3p3-70b-instruct' | 'fireworks-ai:llama4-maverick-instruct-basic' | 'fireworks-ai:llama4-scout-instruct-basic' | 'fireworks-ai:mixtral-8x22b-instruct' | 'fireworks-ai:mixtral-8x7b-instruct' | 'fireworks-ai:mythomax-l2-13b' | 'google-ai:gemini-2.5-flash' | 'google-ai:gemini-2.5-pro' | 'google-ai:models/gemini-2.0-flash' | 'groq:deepseek-r1-distill-llama-70b' | 'groq:gemma2-9b-it' | 'groq:llama-3.1-8b-instant' | 'groq:llama-3.3-70b-versatile' | 'groq:openai/gpt-oss-120b' | 'groq:openai/gpt-oss-20b' | 'openai:gpt-4.1-2025-04-14' | 'openai:gpt-4.1-mini-2025-04-14' | 'openai:gpt-4.1-nano-2025-04-14' | 'openai:gpt-4o-2024-11-20' | 'openai:gpt-4o-mini-2024-07-18' | 'openai:gpt-5-2025-08-07' | 'openai:gpt-5-mini-2025-08-07' | 'openai:gpt-5-nano-2025-08-07' | 'openai:o1-2024-12-17' | 'openai:o1-mini-2024-09-12' | 'openai:o3-2025-04-16' | 'openai:o3-mini-2025-01-31' | 'openai:o4-mini-2025-04-16' | 'openrouter:gpt-oss-120b' | 'xai:grok-3' | 'xai:grok-3-mini' | 'xai:grok-4-0709' | 'xai:grok-4-fast-non-reasoning' | 'xai:grok-4-fast-reasoning' | 'xai:grok-code-fast-1' | ({} & string);
18983
18983
  type CognitiveRequest = {
18984
18984
  /**
18985
18985
  * @minItems 1
@@ -19055,13 +19055,17 @@ type CognitiveResponse = {
19055
19055
  model?: string;
19056
19056
  usage: {
19057
19057
  inputTokens: number;
19058
+ inputCost: number;
19058
19059
  outputTokens: number;
19059
- reasoningTokens?: number;
19060
+ outputCost: number;
19060
19061
  };
19061
19062
  cost?: number;
19062
19063
  cached?: boolean;
19064
+ /**
19065
+ * Time it took for the provider to respond to the LLM query
19066
+ */
19063
19067
  latency?: number;
19064
- stopReason?: string;
19068
+ stopReason?: 'stop' | 'length' | 'content_filter' | 'error';
19065
19069
  reasoningEffort?: string;
19066
19070
  warnings?: {
19067
19071
  type: 'parameter_ignored' | 'provider_limitation' | 'deprecated_model' | 'fallback_used';
package/dist/index.mjs CHANGED
@@ -1311,6 +1311,96 @@ var models = {
1311
1311
  tags: ["low-cost", "general-purpose"],
1312
1312
  lifecycle: "live"
1313
1313
  },
1314
+ "xai:grok-code-fast-1": {
1315
+ id: "xai:grok-code-fast-1",
1316
+ name: "Grok Code Fast 1",
1317
+ description: "Fast coding-optimized Grok model with large context window.",
1318
+ input: {
1319
+ maxTokens: 256e3,
1320
+ costPer1MTokens: 0.2
1321
+ },
1322
+ output: {
1323
+ maxTokens: 32768,
1324
+ costPer1MTokens: 1.5
1325
+ },
1326
+ tags: ["coding", "general-purpose", "low-cost"],
1327
+ lifecycle: "live"
1328
+ },
1329
+ "xai:grok-4-fast-reasoning": {
1330
+ id: "xai:grok-4-fast-reasoning",
1331
+ name: "Grok 4 Fast (Reasoning)",
1332
+ description: "Advanced fast Grok model with reasoning and very large context.",
1333
+ input: {
1334
+ maxTokens: 2e6,
1335
+ costPer1MTokens: 0.2
1336
+ },
1337
+ output: {
1338
+ maxTokens: 128e3,
1339
+ costPer1MTokens: 0.5
1340
+ },
1341
+ tags: ["reasoning", "recommended", "general-purpose"],
1342
+ lifecycle: "live"
1343
+ },
1344
+ "xai:grok-4-fast-non-reasoning": {
1345
+ id: "xai:grok-4-fast-non-reasoning",
1346
+ name: "Grok 4 Fast (Non-Reasoning)",
1347
+ description: "Fast, cost-effective Grok model for non-reasoning tasks.",
1348
+ input: {
1349
+ maxTokens: 2e6,
1350
+ costPer1MTokens: 0.2
1351
+ },
1352
+ output: {
1353
+ maxTokens: 128e3,
1354
+ costPer1MTokens: 0.5
1355
+ },
1356
+ tags: ["low-cost", "recommended", "general-purpose"],
1357
+ lifecycle: "live"
1358
+ },
1359
+ "xai:grok-4-0709": {
1360
+ id: "xai:grok-4-0709",
1361
+ name: "Grok 4 (0709)",
1362
+ description: "Comprehensive Grok 4 model for general-purpose tasks.",
1363
+ input: {
1364
+ maxTokens: 256e3,
1365
+ costPer1MTokens: 3
1366
+ },
1367
+ output: {
1368
+ maxTokens: 32768,
1369
+ costPer1MTokens: 15
1370
+ },
1371
+ tags: ["reasoning", "general-purpose"],
1372
+ lifecycle: "live"
1373
+ },
1374
+ "xai:grok-3-mini": {
1375
+ id: "xai:grok-3-mini",
1376
+ name: "Grok 3 Mini",
1377
+ description: "Lightweight Grok model for cost-sensitive workloads.",
1378
+ input: {
1379
+ maxTokens: 131072,
1380
+ costPer1MTokens: 0.3
1381
+ },
1382
+ output: {
1383
+ maxTokens: 16384,
1384
+ costPer1MTokens: 0.5
1385
+ },
1386
+ tags: ["low-cost", "general-purpose"],
1387
+ lifecycle: "live"
1388
+ },
1389
+ "xai:grok-3": {
1390
+ id: "xai:grok-3",
1391
+ name: "Grok 3",
1392
+ description: "Enterprise-grade Grok model for general-purpose tasks.",
1393
+ input: {
1394
+ maxTokens: 131072,
1395
+ costPer1MTokens: 3
1396
+ },
1397
+ output: {
1398
+ maxTokens: 16384,
1399
+ costPer1MTokens: 15
1400
+ },
1401
+ tags: ["general-purpose"],
1402
+ lifecycle: "live"
1403
+ },
1314
1404
  "openrouter:gpt-oss-120b": {
1315
1405
  id: "openrouter:gpt-oss-120b",
1316
1406
  name: "GPT-OSS 120B (Preview)",
@@ -1326,8 +1416,8 @@ var models = {
1326
1416
  tags: ["preview", "general-purpose", "reasoning"],
1327
1417
  lifecycle: "live"
1328
1418
  },
1329
- "fireworks:gpt-oss-20b": {
1330
- id: "fireworks:gpt-oss-20b",
1419
+ "fireworks-ai:gpt-oss-20b": {
1420
+ id: "fireworks-ai:gpt-oss-20b",
1331
1421
  name: "GPT-OSS 20B",
1332
1422
  description: "gpt-oss-20b is a compact, open-weight language model optimized for low-latency. It shares the same training foundation and capabilities as the GPT-OSS 120B model, with faster responses and lower cost.",
1333
1423
  input: {
@@ -1342,8 +1432,8 @@ var models = {
1342
1432
  lifecycle: "live",
1343
1433
  aliases: ["accounts/fireworks/models/gpt-oss-20b"]
1344
1434
  },
1345
- "fireworks:gpt-oss-120b": {
1346
- id: "fireworks:gpt-oss-120b",
1435
+ "fireworks-ai:gpt-oss-120b": {
1436
+ id: "fireworks-ai:gpt-oss-120b",
1347
1437
  name: "GPT-OSS 120B",
1348
1438
  description: "gpt-oss-120b is a high-performance, open-weight language model designed for production-grade, general-purpose use cases. It excels at complex reasoning and supports configurable reasoning effort, full chain-of-thought transparency for easier debugging and trust, and native agentic capabilities for function calling, tool use, and structured outputs.",
1349
1439
  input: {
@@ -1358,8 +1448,8 @@ var models = {
1358
1448
  lifecycle: "live",
1359
1449
  aliases: ["accounts/fireworks/models/gpt-oss-120b"]
1360
1450
  },
1361
- "fireworks:deepseek-r1-0528": {
1362
- id: "fireworks:deepseek-r1-0528",
1451
+ "fireworks-ai:deepseek-r1-0528": {
1452
+ id: "fireworks-ai:deepseek-r1-0528",
1363
1453
  name: "DeepSeek R1 0528",
1364
1454
  description: "The updated DeepSeek R1 0528 model delivers major improvements in reasoning, inference, and accuracy through enhanced post-training optimization and greater computational resources. It now performs at a level approaching top-tier models like OpenAI o3 and Gemini 2.5 Pro, with notable gains in complex tasks such as math and programming. The update also reduces hallucinations, improves function calling, and enhances the coding experience.",
1365
1455
  input: {
@@ -1374,8 +1464,8 @@ var models = {
1374
1464
  lifecycle: "live",
1375
1465
  aliases: ["accounts/fireworks/models/deepseek-r1-0528"]
1376
1466
  },
1377
- "fireworks:deepseek-v3-0324": {
1378
- id: "fireworks:deepseek-v3-0324",
1467
+ "fireworks-ai:deepseek-v3-0324": {
1468
+ id: "fireworks-ai:deepseek-v3-0324",
1379
1469
  name: "DeepSeek V3 0324",
1380
1470
  description: "DeepSeek V3, a 685B-parameter, mixture-of-experts model, is the latest iteration of the flagship chat model family from the DeepSeek team. It succeeds the DeepSeek V3 model and performs really well on a variety of tasks.",
1381
1471
  input: {
@@ -1390,8 +1480,8 @@ var models = {
1390
1480
  lifecycle: "live",
1391
1481
  aliases: ["accounts/fireworks/models/deepseek-v3-0324"]
1392
1482
  },
1393
- "fireworks:llama4-maverick-instruct-basic": {
1394
- id: "fireworks:llama4-maverick-instruct-basic",
1483
+ "fireworks-ai:llama4-maverick-instruct-basic": {
1484
+ id: "fireworks-ai:llama4-maverick-instruct-basic",
1395
1485
  name: "Llama 4 Maverick Instruct (Basic)",
1396
1486
  description: "Llama 4 Maverick 17B Instruct (128E) is a high-capacity multimodal language model from Meta, built on a mixture-of-experts (MoE) architecture with 128 experts and 17 billion active parameters per forward pass (400B total). It supports multilingual text and image input, and produces multilingual text and code output across 12 supported languages. Optimized for vision-language tasks, Maverick is instruction-tuned for assistant-like behavior, image reasoning, and general-purpose multimodal interaction, and suited for research and commercial applications requiring advanced multimodal understanding and high model throughput.",
1397
1487
  input: {
@@ -1406,8 +1496,8 @@ var models = {
1406
1496
  lifecycle: "live",
1407
1497
  aliases: ["accounts/fireworks/models/llama4-maverick-instruct-basic"]
1408
1498
  },
1409
- "fireworks:llama4-scout-instruct-basic": {
1410
- id: "fireworks:llama4-scout-instruct-basic",
1499
+ "fireworks-ai:llama4-scout-instruct-basic": {
1500
+ id: "fireworks-ai:llama4-scout-instruct-basic",
1411
1501
  name: "Llama 4 Scout Instruct (Basic)",
1412
1502
  description: "Llama 4 Scout 17B Instruct (16E) is a mixture-of-experts (MoE) language model developed by Meta, uses 16 experts per forward pass, activating 17 billion parameters out of a total of 109B. It supports native multimodal input (text and image) and multilingual output (text and code) across 12 supported languages. Designed for assistant-style interaction and visual reasoning, it is instruction-tuned for use in multilingual chat, captioning, and image understanding tasks.",
1413
1503
  input: {
@@ -1422,8 +1512,8 @@ var models = {
1422
1512
  lifecycle: "live",
1423
1513
  aliases: ["accounts/fireworks/models/llama4-scout-instruct-basic"]
1424
1514
  },
1425
- "fireworks:llama-v3p3-70b-instruct": {
1426
- id: "fireworks:llama-v3p3-70b-instruct",
1515
+ "fireworks-ai:llama-v3p3-70b-instruct": {
1516
+ id: "fireworks-ai:llama-v3p3-70b-instruct",
1427
1517
  name: "Llama 3.3 70B Instruct",
1428
1518
  description: "Llama 3.3 70B Instruct is the December update of Llama 3.1 70B. The model improves upon Llama 3.1 70B (released July 2024) with advances in tool calling, multilingual text support, math and coding. The model achieves industry leading results in reasoning, math and instruction following and provides similar performance as 3.1 405B but with significant speed and cost improvements.",
1429
1519
  input: {
@@ -1438,8 +1528,8 @@ var models = {
1438
1528
  lifecycle: "live",
1439
1529
  aliases: ["accounts/fireworks/models/llama-v3p3-70b-instruct"]
1440
1530
  },
1441
- "fireworks:deepseek-r1": {
1442
- id: "fireworks:deepseek-r1",
1531
+ "fireworks-ai:deepseek-r1": {
1532
+ id: "fireworks-ai:deepseek-r1",
1443
1533
  name: "DeepSeek R1 (Fast)",
1444
1534
  description: "This version of the R1 model has a perfect balance between speed and cost-efficiency for real-time interactive experiences, with speeds up to 90 tokens per second.\n\nDeepSeek-R1 is a state-of-the-art large language model optimized with reinforcement learning and cold-start data for exceptional reasoning, math, and code performance. **Note**: This model will always use a temperature of 0.6 as recommended by DeepSeek.",
1445
1535
  input: {
@@ -1454,8 +1544,8 @@ var models = {
1454
1544
  lifecycle: "live",
1455
1545
  aliases: ["accounts/fireworks/models/deepseek-r1"]
1456
1546
  },
1457
- "fireworks:deepseek-r1-basic": {
1458
- id: "fireworks:deepseek-r1-basic",
1547
+ "fireworks-ai:deepseek-r1-basic": {
1548
+ id: "fireworks-ai:deepseek-r1-basic",
1459
1549
  name: "DeepSeek R1 (Basic)",
1460
1550
  description: 'This version of the R1 model is optimized for throughput and cost-effectiveness and has a lower cost but slightly higher latency than the "Fast" version of the model.\n\nDeepSeek-R1 is a state-of-the-art large language model optimized with reinforcement learning and cold-start data for exceptional reasoning, math, and code performance. **Note**: This model will always use a temperature of 0.6 as recommended by DeepSeek.',
1461
1551
  input: {
@@ -1470,8 +1560,8 @@ var models = {
1470
1560
  lifecycle: "live",
1471
1561
  aliases: ["accounts/fireworks/models/deepseek-r1-basic"]
1472
1562
  },
1473
- "fireworks:deepseek-v3": {
1474
- id: "fireworks:deepseek-v3",
1563
+ "fireworks-ai:deepseek-v3": {
1564
+ id: "fireworks-ai:deepseek-v3",
1475
1565
  name: "DeepSeek V3",
1476
1566
  description: "A a strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek.",
1477
1567
  input: {
@@ -1486,8 +1576,8 @@ var models = {
1486
1576
  lifecycle: "deprecated",
1487
1577
  aliases: ["accounts/fireworks/models/deepseek-v3"]
1488
1578
  },
1489
- "fireworks:llama-v3p1-405b-instruct": {
1490
- id: "fireworks:llama-v3p1-405b-instruct",
1579
+ "fireworks-ai:llama-v3p1-405b-instruct": {
1580
+ id: "fireworks-ai:llama-v3p1-405b-instruct",
1491
1581
  name: "Llama 3.1 405B Instruct",
1492
1582
  description: "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models in 8B, 70B and 405B sizes. The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",
1493
1583
  input: {
@@ -1502,8 +1592,8 @@ var models = {
1502
1592
  lifecycle: "deprecated",
1503
1593
  aliases: ["accounts/fireworks/models/llama-v3p1-405b-instruct"]
1504
1594
  },
1505
- "fireworks:llama-v3p1-70b-instruct": {
1506
- id: "fireworks:llama-v3p1-70b-instruct",
1595
+ "fireworks-ai:llama-v3p1-70b-instruct": {
1596
+ id: "fireworks-ai:llama-v3p1-70b-instruct",
1507
1597
  name: "Llama 3.1 70B Instruct",
1508
1598
  description: "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models in 8B, 70B and 405B sizes. The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",
1509
1599
  input: {
@@ -1518,8 +1608,8 @@ var models = {
1518
1608
  lifecycle: "deprecated",
1519
1609
  aliases: ["accounts/fireworks/models/llama-v3p1-70b-instruct"]
1520
1610
  },
1521
- "fireworks:llama-v3p1-8b-instruct": {
1522
- id: "fireworks:llama-v3p1-8b-instruct",
1611
+ "fireworks-ai:llama-v3p1-8b-instruct": {
1612
+ id: "fireworks-ai:llama-v3p1-8b-instruct",
1523
1613
  name: "Llama 3.1 8B Instruct",
1524
1614
  description: "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models in 8B, 70B and 405B sizes. The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",
1525
1615
  input: {
@@ -1534,8 +1624,8 @@ var models = {
1534
1624
  lifecycle: "live",
1535
1625
  aliases: ["accounts/fireworks/models/llama-v3p1-8b-instruct"]
1536
1626
  },
1537
- "fireworks:mixtral-8x22b-instruct": {
1538
- id: "fireworks:mixtral-8x22b-instruct",
1627
+ "fireworks-ai:mixtral-8x22b-instruct": {
1628
+ id: "fireworks-ai:mixtral-8x22b-instruct",
1539
1629
  name: "Mixtral MoE 8x22B Instruct",
1540
1630
  description: "Mistral MoE 8x22B Instruct v0.1 model with Sparse Mixture of Experts. Fine tuned for instruction following.",
1541
1631
  input: {
@@ -1550,8 +1640,8 @@ var models = {
1550
1640
  lifecycle: "live",
1551
1641
  aliases: ["accounts/fireworks/models/mixtral-8x22b-instruct"]
1552
1642
  },
1553
- "fireworks:mixtral-8x7b-instruct": {
1554
- id: "fireworks:mixtral-8x7b-instruct",
1643
+ "fireworks-ai:mixtral-8x7b-instruct": {
1644
+ id: "fireworks-ai:mixtral-8x7b-instruct",
1555
1645
  name: "Mixtral MoE 8x7B Instruct",
1556
1646
  description: "Mistral MoE 8x7B Instruct v0.1 model with Sparse Mixture of Experts. Fine tuned for instruction following",
1557
1647
  input: {
@@ -1566,8 +1656,8 @@ var models = {
1566
1656
  lifecycle: "live",
1567
1657
  aliases: ["accounts/fireworks/models/mixtral-8x7b-instruct"]
1568
1658
  },
1569
- "fireworks:mythomax-l2-13b": {
1570
- id: "fireworks:mythomax-l2-13b",
1659
+ "fireworks-ai:mythomax-l2-13b": {
1660
+ id: "fireworks-ai:mythomax-l2-13b",
1571
1661
  name: "MythoMax L2 13b",
1572
1662
  description: "MythoMax L2 is designed to excel at both roleplaying and storytelling, and is an improved variant of the previous MythoMix model, combining the MythoLogic-L2 and Huginn models.",
1573
1663
  input: {
@@ -1582,8 +1672,8 @@ var models = {
1582
1672
  lifecycle: "live",
1583
1673
  aliases: ["accounts/fireworks/models/mythomax-l2-13b"]
1584
1674
  },
1585
- "fireworks:gemma2-9b-it": {
1586
- id: "fireworks:gemma2-9b-it",
1675
+ "fireworks-ai:gemma2-9b-it": {
1676
+ id: "fireworks-ai:gemma2-9b-it",
1587
1677
  name: "Gemma 2 9B Instruct",
1588
1678
  description: "Redesigned for outsized performance and unmatched efficiency, Gemma 2 optimizes for blazing-fast inference on diverse hardware. Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning.",
1589
1679
  input: {
@@ -1717,7 +1807,7 @@ var CognitiveBeta = class {
1717
1807
  }
1718
1808
  const res = await this._withServerRetry(
1719
1809
  () => this._axiosClient.post(
1720
- "/v1/generate-text-stream",
1810
+ "/v2/cognitive/generate-text-stream",
1721
1811
  { ...request, stream: true },
1722
1812
  {
1723
1813
  responseType: "stream",