@botpress/cognitive 0.1.42 → 0.1.44
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.turbo/turbo-build.log +7 -7
- package/dist/index.cjs +125 -35
- package/dist/index.cjs.map +2 -2
- package/dist/index.d.ts +7 -3
- package/dist/index.mjs +125 -35
- package/dist/index.mjs.map +2 -2
- package/package.json +1 -1
package/dist/index.d.ts
CHANGED
|
@@ -18979,7 +18979,7 @@ declare class Cognitive {
|
|
|
18979
18979
|
private _generateContent;
|
|
18980
18980
|
}
|
|
18981
18981
|
|
|
18982
|
-
type Models = 'auto' | 'best' | 'fast' | 'reasoning' | 'cheapest' | 'balance' | 'recommended' | 'reasoning' | 'general-purpose' | 'low-cost' | 'vision' | 'coding' | 'function-calling' | 'agents' | 'storytelling' | 'preview' | 'roleplay' | 'anthropic:claude-3-5-haiku-20241022' | 'anthropic:claude-3-5-sonnet-20240620' | 'anthropic:claude-3-5-sonnet-20241022' | 'anthropic:claude-3-7-sonnet-20250219' | 'anthropic:claude-3-haiku-20240307' | 'anthropic:claude-sonnet-4-20250514' | 'cerebras:gpt-oss-120b' | 'cerebras:llama-4-scout-17b-16e-instruct' | 'cerebras:llama3.1-8b' | 'cerebras:llama3.3-70b' | 'cerebras:qwen-3-32b' | 'fireworks:deepseek-r1' | 'fireworks:deepseek-r1-0528' | 'fireworks:deepseek-r1-basic' | 'fireworks:deepseek-v3-0324' | 'fireworks:gpt-oss-120b' | 'fireworks:gpt-oss-20b' | 'fireworks:llama-v3p1-8b-instruct' | 'fireworks:llama-v3p3-70b-instruct' | 'fireworks:llama4-maverick-instruct-basic' | 'fireworks:llama4-scout-instruct-basic' | 'fireworks:mixtral-8x22b-instruct' | 'fireworks:mixtral-8x7b-instruct' | 'fireworks:mythomax-l2-13b' | 'google-ai:gemini-2.5-flash' | 'google-ai:gemini-2.5-pro' | 'google-ai:models/gemini-2.0-flash' | 'groq:deepseek-r1-distill-llama-70b' | 'groq:gemma2-9b-it' | 'groq:llama-3.1-8b-instant' | 'groq:llama-3.3-70b-versatile' | 'groq:openai/gpt-oss-120b' | 'groq:openai/gpt-oss-20b' | 'openai:gpt-4.1-2025-04-14' | 'openai:gpt-4.1-mini-2025-04-14' | 'openai:gpt-4.1-nano-2025-04-14' | 'openai:gpt-4o-2024-11-20' | 'openai:gpt-4o-mini-2024-07-18' | 'openai:gpt-5-2025-08-07' | 'openai:gpt-5-mini-2025-08-07' | 'openai:gpt-5-nano-2025-08-07' | 'openai:o1-2024-12-17' | 'openai:o1-mini-2024-09-12' | 'openai:o3-2025-04-16' | 'openai:o3-mini-2025-01-31' | 'openai:o4-mini-2025-04-16' | 'openrouter:gpt-oss-120b' | ({} & string);
|
|
18982
|
+
type Models = 'auto' | 'best' | 'fast' | 'reasoning' | 'cheapest' | 'balance' | 'recommended' | 'reasoning' | 'general-purpose' | 'low-cost' | 'vision' | 'coding' | 'function-calling' | 'agents' | 'storytelling' | 'preview' | 'roleplay' | 'anthropic:claude-3-5-haiku-20241022' | 'anthropic:claude-3-5-sonnet-20240620' | 'anthropic:claude-3-5-sonnet-20241022' | 'anthropic:claude-3-7-sonnet-20250219' | 'anthropic:claude-3-haiku-20240307' | 'anthropic:claude-sonnet-4-20250514' | 'cerebras:gpt-oss-120b' | 'cerebras:llama-4-scout-17b-16e-instruct' | 'cerebras:llama3.1-8b' | 'cerebras:llama3.3-70b' | 'cerebras:qwen-3-32b' | 'fireworks-ai:deepseek-r1' | 'fireworks-ai:deepseek-r1-0528' | 'fireworks-ai:deepseek-r1-basic' | 'fireworks-ai:deepseek-v3-0324' | 'fireworks-ai:gpt-oss-120b' | 'fireworks-ai:gpt-oss-20b' | 'fireworks-ai:llama-v3p1-8b-instruct' | 'fireworks-ai:llama-v3p3-70b-instruct' | 'fireworks-ai:llama4-maverick-instruct-basic' | 'fireworks-ai:llama4-scout-instruct-basic' | 'fireworks-ai:mixtral-8x22b-instruct' | 'fireworks-ai:mixtral-8x7b-instruct' | 'fireworks-ai:mythomax-l2-13b' | 'google-ai:gemini-2.5-flash' | 'google-ai:gemini-2.5-pro' | 'google-ai:models/gemini-2.0-flash' | 'groq:deepseek-r1-distill-llama-70b' | 'groq:gemma2-9b-it' | 'groq:llama-3.1-8b-instant' | 'groq:llama-3.3-70b-versatile' | 'groq:openai/gpt-oss-120b' | 'groq:openai/gpt-oss-20b' | 'openai:gpt-4.1-2025-04-14' | 'openai:gpt-4.1-mini-2025-04-14' | 'openai:gpt-4.1-nano-2025-04-14' | 'openai:gpt-4o-2024-11-20' | 'openai:gpt-4o-mini-2024-07-18' | 'openai:gpt-5-2025-08-07' | 'openai:gpt-5-mini-2025-08-07' | 'openai:gpt-5-nano-2025-08-07' | 'openai:o1-2024-12-17' | 'openai:o1-mini-2024-09-12' | 'openai:o3-2025-04-16' | 'openai:o3-mini-2025-01-31' | 'openai:o4-mini-2025-04-16' | 'openrouter:gpt-oss-120b' | 'xai:grok-3' | 'xai:grok-3-mini' | 'xai:grok-4-0709' | 'xai:grok-4-fast-non-reasoning' | 'xai:grok-4-fast-reasoning' | 'xai:grok-code-fast-1' | ({} & string);
|
|
18983
18983
|
type CognitiveRequest = {
|
|
18984
18984
|
/**
|
|
18985
18985
|
* @minItems 1
|
|
@@ -19055,13 +19055,17 @@ type CognitiveResponse = {
|
|
|
19055
19055
|
model?: string;
|
|
19056
19056
|
usage: {
|
|
19057
19057
|
inputTokens: number;
|
|
19058
|
+
inputCost: number;
|
|
19058
19059
|
outputTokens: number;
|
|
19059
|
-
|
|
19060
|
+
outputCost: number;
|
|
19060
19061
|
};
|
|
19061
19062
|
cost?: number;
|
|
19062
19063
|
cached?: boolean;
|
|
19064
|
+
/**
|
|
19065
|
+
* Time it took for the provider to respond to the LLM query
|
|
19066
|
+
*/
|
|
19063
19067
|
latency?: number;
|
|
19064
|
-
stopReason?:
|
|
19068
|
+
stopReason?: 'stop' | 'length' | 'content_filter' | 'error';
|
|
19065
19069
|
reasoningEffort?: string;
|
|
19066
19070
|
warnings?: {
|
|
19067
19071
|
type: 'parameter_ignored' | 'provider_limitation' | 'deprecated_model' | 'fallback_used';
|
package/dist/index.mjs
CHANGED
|
@@ -1311,6 +1311,96 @@ var models = {
|
|
|
1311
1311
|
tags: ["low-cost", "general-purpose"],
|
|
1312
1312
|
lifecycle: "live"
|
|
1313
1313
|
},
|
|
1314
|
+
"xai:grok-code-fast-1": {
|
|
1315
|
+
id: "xai:grok-code-fast-1",
|
|
1316
|
+
name: "Grok Code Fast 1",
|
|
1317
|
+
description: "Fast coding-optimized Grok model with large context window.",
|
|
1318
|
+
input: {
|
|
1319
|
+
maxTokens: 256e3,
|
|
1320
|
+
costPer1MTokens: 0.2
|
|
1321
|
+
},
|
|
1322
|
+
output: {
|
|
1323
|
+
maxTokens: 32768,
|
|
1324
|
+
costPer1MTokens: 1.5
|
|
1325
|
+
},
|
|
1326
|
+
tags: ["coding", "general-purpose", "low-cost"],
|
|
1327
|
+
lifecycle: "live"
|
|
1328
|
+
},
|
|
1329
|
+
"xai:grok-4-fast-reasoning": {
|
|
1330
|
+
id: "xai:grok-4-fast-reasoning",
|
|
1331
|
+
name: "Grok 4 Fast (Reasoning)",
|
|
1332
|
+
description: "Advanced fast Grok model with reasoning and very large context.",
|
|
1333
|
+
input: {
|
|
1334
|
+
maxTokens: 2e6,
|
|
1335
|
+
costPer1MTokens: 0.2
|
|
1336
|
+
},
|
|
1337
|
+
output: {
|
|
1338
|
+
maxTokens: 128e3,
|
|
1339
|
+
costPer1MTokens: 0.5
|
|
1340
|
+
},
|
|
1341
|
+
tags: ["reasoning", "recommended", "general-purpose"],
|
|
1342
|
+
lifecycle: "live"
|
|
1343
|
+
},
|
|
1344
|
+
"xai:grok-4-fast-non-reasoning": {
|
|
1345
|
+
id: "xai:grok-4-fast-non-reasoning",
|
|
1346
|
+
name: "Grok 4 Fast (Non-Reasoning)",
|
|
1347
|
+
description: "Fast, cost-effective Grok model for non-reasoning tasks.",
|
|
1348
|
+
input: {
|
|
1349
|
+
maxTokens: 2e6,
|
|
1350
|
+
costPer1MTokens: 0.2
|
|
1351
|
+
},
|
|
1352
|
+
output: {
|
|
1353
|
+
maxTokens: 128e3,
|
|
1354
|
+
costPer1MTokens: 0.5
|
|
1355
|
+
},
|
|
1356
|
+
tags: ["low-cost", "recommended", "general-purpose"],
|
|
1357
|
+
lifecycle: "live"
|
|
1358
|
+
},
|
|
1359
|
+
"xai:grok-4-0709": {
|
|
1360
|
+
id: "xai:grok-4-0709",
|
|
1361
|
+
name: "Grok 4 (0709)",
|
|
1362
|
+
description: "Comprehensive Grok 4 model for general-purpose tasks.",
|
|
1363
|
+
input: {
|
|
1364
|
+
maxTokens: 256e3,
|
|
1365
|
+
costPer1MTokens: 3
|
|
1366
|
+
},
|
|
1367
|
+
output: {
|
|
1368
|
+
maxTokens: 32768,
|
|
1369
|
+
costPer1MTokens: 15
|
|
1370
|
+
},
|
|
1371
|
+
tags: ["reasoning", "general-purpose"],
|
|
1372
|
+
lifecycle: "live"
|
|
1373
|
+
},
|
|
1374
|
+
"xai:grok-3-mini": {
|
|
1375
|
+
id: "xai:grok-3-mini",
|
|
1376
|
+
name: "Grok 3 Mini",
|
|
1377
|
+
description: "Lightweight Grok model for cost-sensitive workloads.",
|
|
1378
|
+
input: {
|
|
1379
|
+
maxTokens: 131072,
|
|
1380
|
+
costPer1MTokens: 0.3
|
|
1381
|
+
},
|
|
1382
|
+
output: {
|
|
1383
|
+
maxTokens: 16384,
|
|
1384
|
+
costPer1MTokens: 0.5
|
|
1385
|
+
},
|
|
1386
|
+
tags: ["low-cost", "general-purpose"],
|
|
1387
|
+
lifecycle: "live"
|
|
1388
|
+
},
|
|
1389
|
+
"xai:grok-3": {
|
|
1390
|
+
id: "xai:grok-3",
|
|
1391
|
+
name: "Grok 3",
|
|
1392
|
+
description: "Enterprise-grade Grok model for general-purpose tasks.",
|
|
1393
|
+
input: {
|
|
1394
|
+
maxTokens: 131072,
|
|
1395
|
+
costPer1MTokens: 3
|
|
1396
|
+
},
|
|
1397
|
+
output: {
|
|
1398
|
+
maxTokens: 16384,
|
|
1399
|
+
costPer1MTokens: 15
|
|
1400
|
+
},
|
|
1401
|
+
tags: ["general-purpose"],
|
|
1402
|
+
lifecycle: "live"
|
|
1403
|
+
},
|
|
1314
1404
|
"openrouter:gpt-oss-120b": {
|
|
1315
1405
|
id: "openrouter:gpt-oss-120b",
|
|
1316
1406
|
name: "GPT-OSS 120B (Preview)",
|
|
@@ -1326,8 +1416,8 @@ var models = {
|
|
|
1326
1416
|
tags: ["preview", "general-purpose", "reasoning"],
|
|
1327
1417
|
lifecycle: "live"
|
|
1328
1418
|
},
|
|
1329
|
-
"fireworks:gpt-oss-20b": {
|
|
1330
|
-
id: "fireworks:gpt-oss-20b",
|
|
1419
|
+
"fireworks-ai:gpt-oss-20b": {
|
|
1420
|
+
id: "fireworks-ai:gpt-oss-20b",
|
|
1331
1421
|
name: "GPT-OSS 20B",
|
|
1332
1422
|
description: "gpt-oss-20b is a compact, open-weight language model optimized for low-latency. It shares the same training foundation and capabilities as the GPT-OSS 120B model, with faster responses and lower cost.",
|
|
1333
1423
|
input: {
|
|
@@ -1342,8 +1432,8 @@ var models = {
|
|
|
1342
1432
|
lifecycle: "live",
|
|
1343
1433
|
aliases: ["accounts/fireworks/models/gpt-oss-20b"]
|
|
1344
1434
|
},
|
|
1345
|
-
"fireworks:gpt-oss-120b": {
|
|
1346
|
-
id: "fireworks:gpt-oss-120b",
|
|
1435
|
+
"fireworks-ai:gpt-oss-120b": {
|
|
1436
|
+
id: "fireworks-ai:gpt-oss-120b",
|
|
1347
1437
|
name: "GPT-OSS 120B",
|
|
1348
1438
|
description: "gpt-oss-120b is a high-performance, open-weight language model designed for production-grade, general-purpose use cases. It excels at complex reasoning and supports configurable reasoning effort, full chain-of-thought transparency for easier debugging and trust, and native agentic capabilities for function calling, tool use, and structured outputs.",
|
|
1349
1439
|
input: {
|
|
@@ -1358,8 +1448,8 @@ var models = {
|
|
|
1358
1448
|
lifecycle: "live",
|
|
1359
1449
|
aliases: ["accounts/fireworks/models/gpt-oss-120b"]
|
|
1360
1450
|
},
|
|
1361
|
-
"fireworks:deepseek-r1-0528": {
|
|
1362
|
-
id: "fireworks:deepseek-r1-0528",
|
|
1451
|
+
"fireworks-ai:deepseek-r1-0528": {
|
|
1452
|
+
id: "fireworks-ai:deepseek-r1-0528",
|
|
1363
1453
|
name: "DeepSeek R1 0528",
|
|
1364
1454
|
description: "The updated DeepSeek R1 0528 model delivers major improvements in reasoning, inference, and accuracy through enhanced post-training optimization and greater computational resources. It now performs at a level approaching top-tier models like OpenAI o3 and Gemini 2.5 Pro, with notable gains in complex tasks such as math and programming. The update also reduces hallucinations, improves function calling, and enhances the coding experience.",
|
|
1365
1455
|
input: {
|
|
@@ -1374,8 +1464,8 @@ var models = {
|
|
|
1374
1464
|
lifecycle: "live",
|
|
1375
1465
|
aliases: ["accounts/fireworks/models/deepseek-r1-0528"]
|
|
1376
1466
|
},
|
|
1377
|
-
"fireworks:deepseek-v3-0324": {
|
|
1378
|
-
id: "fireworks:deepseek-v3-0324",
|
|
1467
|
+
"fireworks-ai:deepseek-v3-0324": {
|
|
1468
|
+
id: "fireworks-ai:deepseek-v3-0324",
|
|
1379
1469
|
name: "DeepSeek V3 0324",
|
|
1380
1470
|
description: "DeepSeek V3, a 685B-parameter, mixture-of-experts model, is the latest iteration of the flagship chat model family from the DeepSeek team. It succeeds the DeepSeek V3 model and performs really well on a variety of tasks.",
|
|
1381
1471
|
input: {
|
|
@@ -1390,8 +1480,8 @@ var models = {
|
|
|
1390
1480
|
lifecycle: "live",
|
|
1391
1481
|
aliases: ["accounts/fireworks/models/deepseek-v3-0324"]
|
|
1392
1482
|
},
|
|
1393
|
-
"fireworks:llama4-maverick-instruct-basic": {
|
|
1394
|
-
id: "fireworks:llama4-maverick-instruct-basic",
|
|
1483
|
+
"fireworks-ai:llama4-maverick-instruct-basic": {
|
|
1484
|
+
id: "fireworks-ai:llama4-maverick-instruct-basic",
|
|
1395
1485
|
name: "Llama 4 Maverick Instruct (Basic)",
|
|
1396
1486
|
description: "Llama 4 Maverick 17B Instruct (128E) is a high-capacity multimodal language model from Meta, built on a mixture-of-experts (MoE) architecture with 128 experts and 17 billion active parameters per forward pass (400B total). It supports multilingual text and image input, and produces multilingual text and code output across 12 supported languages. Optimized for vision-language tasks, Maverick is instruction-tuned for assistant-like behavior, image reasoning, and general-purpose multimodal interaction, and suited for research and commercial applications requiring advanced multimodal understanding and high model throughput.",
|
|
1397
1487
|
input: {
|
|
@@ -1406,8 +1496,8 @@ var models = {
|
|
|
1406
1496
|
lifecycle: "live",
|
|
1407
1497
|
aliases: ["accounts/fireworks/models/llama4-maverick-instruct-basic"]
|
|
1408
1498
|
},
|
|
1409
|
-
"fireworks:llama4-scout-instruct-basic": {
|
|
1410
|
-
id: "fireworks:llama4-scout-instruct-basic",
|
|
1499
|
+
"fireworks-ai:llama4-scout-instruct-basic": {
|
|
1500
|
+
id: "fireworks-ai:llama4-scout-instruct-basic",
|
|
1411
1501
|
name: "Llama 4 Scout Instruct (Basic)",
|
|
1412
1502
|
description: "Llama 4 Scout 17B Instruct (16E) is a mixture-of-experts (MoE) language model developed by Meta, uses 16 experts per forward pass, activating 17 billion parameters out of a total of 109B. It supports native multimodal input (text and image) and multilingual output (text and code) across 12 supported languages. Designed for assistant-style interaction and visual reasoning, it is instruction-tuned for use in multilingual chat, captioning, and image understanding tasks.",
|
|
1413
1503
|
input: {
|
|
@@ -1422,8 +1512,8 @@ var models = {
|
|
|
1422
1512
|
lifecycle: "live",
|
|
1423
1513
|
aliases: ["accounts/fireworks/models/llama4-scout-instruct-basic"]
|
|
1424
1514
|
},
|
|
1425
|
-
"fireworks:llama-v3p3-70b-instruct": {
|
|
1426
|
-
id: "fireworks:llama-v3p3-70b-instruct",
|
|
1515
|
+
"fireworks-ai:llama-v3p3-70b-instruct": {
|
|
1516
|
+
id: "fireworks-ai:llama-v3p3-70b-instruct",
|
|
1427
1517
|
name: "Llama 3.3 70B Instruct",
|
|
1428
1518
|
description: "Llama 3.3 70B Instruct is the December update of Llama 3.1 70B. The model improves upon Llama 3.1 70B (released July 2024) with advances in tool calling, multilingual text support, math and coding. The model achieves industry leading results in reasoning, math and instruction following and provides similar performance as 3.1 405B but with significant speed and cost improvements.",
|
|
1429
1519
|
input: {
|
|
@@ -1438,8 +1528,8 @@ var models = {
|
|
|
1438
1528
|
lifecycle: "live",
|
|
1439
1529
|
aliases: ["accounts/fireworks/models/llama-v3p3-70b-instruct"]
|
|
1440
1530
|
},
|
|
1441
|
-
"fireworks:deepseek-r1": {
|
|
1442
|
-
id: "fireworks:deepseek-r1",
|
|
1531
|
+
"fireworks-ai:deepseek-r1": {
|
|
1532
|
+
id: "fireworks-ai:deepseek-r1",
|
|
1443
1533
|
name: "DeepSeek R1 (Fast)",
|
|
1444
1534
|
description: "This version of the R1 model has a perfect balance between speed and cost-efficiency for real-time interactive experiences, with speeds up to 90 tokens per second.\n\nDeepSeek-R1 is a state-of-the-art large language model optimized with reinforcement learning and cold-start data for exceptional reasoning, math, and code performance. **Note**: This model will always use a temperature of 0.6 as recommended by DeepSeek.",
|
|
1445
1535
|
input: {
|
|
@@ -1454,8 +1544,8 @@ var models = {
|
|
|
1454
1544
|
lifecycle: "live",
|
|
1455
1545
|
aliases: ["accounts/fireworks/models/deepseek-r1"]
|
|
1456
1546
|
},
|
|
1457
|
-
"fireworks:deepseek-r1-basic": {
|
|
1458
|
-
id: "fireworks:deepseek-r1-basic",
|
|
1547
|
+
"fireworks-ai:deepseek-r1-basic": {
|
|
1548
|
+
id: "fireworks-ai:deepseek-r1-basic",
|
|
1459
1549
|
name: "DeepSeek R1 (Basic)",
|
|
1460
1550
|
description: 'This version of the R1 model is optimized for throughput and cost-effectiveness and has a lower cost but slightly higher latency than the "Fast" version of the model.\n\nDeepSeek-R1 is a state-of-the-art large language model optimized with reinforcement learning and cold-start data for exceptional reasoning, math, and code performance. **Note**: This model will always use a temperature of 0.6 as recommended by DeepSeek.',
|
|
1461
1551
|
input: {
|
|
@@ -1470,8 +1560,8 @@ var models = {
|
|
|
1470
1560
|
lifecycle: "live",
|
|
1471
1561
|
aliases: ["accounts/fireworks/models/deepseek-r1-basic"]
|
|
1472
1562
|
},
|
|
1473
|
-
"fireworks:deepseek-v3": {
|
|
1474
|
-
id: "fireworks:deepseek-v3",
|
|
1563
|
+
"fireworks-ai:deepseek-v3": {
|
|
1564
|
+
id: "fireworks-ai:deepseek-v3",
|
|
1475
1565
|
name: "DeepSeek V3",
|
|
1476
1566
|
description: "A a strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek.",
|
|
1477
1567
|
input: {
|
|
@@ -1486,8 +1576,8 @@ var models = {
|
|
|
1486
1576
|
lifecycle: "deprecated",
|
|
1487
1577
|
aliases: ["accounts/fireworks/models/deepseek-v3"]
|
|
1488
1578
|
},
|
|
1489
|
-
"fireworks:llama-v3p1-405b-instruct": {
|
|
1490
|
-
id: "fireworks:llama-v3p1-405b-instruct",
|
|
1579
|
+
"fireworks-ai:llama-v3p1-405b-instruct": {
|
|
1580
|
+
id: "fireworks-ai:llama-v3p1-405b-instruct",
|
|
1491
1581
|
name: "Llama 3.1 405B Instruct",
|
|
1492
1582
|
description: "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models in 8B, 70B and 405B sizes. The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",
|
|
1493
1583
|
input: {
|
|
@@ -1502,8 +1592,8 @@ var models = {
|
|
|
1502
1592
|
lifecycle: "deprecated",
|
|
1503
1593
|
aliases: ["accounts/fireworks/models/llama-v3p1-405b-instruct"]
|
|
1504
1594
|
},
|
|
1505
|
-
"fireworks:llama-v3p1-70b-instruct": {
|
|
1506
|
-
id: "fireworks:llama-v3p1-70b-instruct",
|
|
1595
|
+
"fireworks-ai:llama-v3p1-70b-instruct": {
|
|
1596
|
+
id: "fireworks-ai:llama-v3p1-70b-instruct",
|
|
1507
1597
|
name: "Llama 3.1 70B Instruct",
|
|
1508
1598
|
description: "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models in 8B, 70B and 405B sizes. The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",
|
|
1509
1599
|
input: {
|
|
@@ -1518,8 +1608,8 @@ var models = {
|
|
|
1518
1608
|
lifecycle: "deprecated",
|
|
1519
1609
|
aliases: ["accounts/fireworks/models/llama-v3p1-70b-instruct"]
|
|
1520
1610
|
},
|
|
1521
|
-
"fireworks:llama-v3p1-8b-instruct": {
|
|
1522
|
-
id: "fireworks:llama-v3p1-8b-instruct",
|
|
1611
|
+
"fireworks-ai:llama-v3p1-8b-instruct": {
|
|
1612
|
+
id: "fireworks-ai:llama-v3p1-8b-instruct",
|
|
1523
1613
|
name: "Llama 3.1 8B Instruct",
|
|
1524
1614
|
description: "The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models in 8B, 70B and 405B sizes. The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.",
|
|
1525
1615
|
input: {
|
|
@@ -1534,8 +1624,8 @@ var models = {
|
|
|
1534
1624
|
lifecycle: "live",
|
|
1535
1625
|
aliases: ["accounts/fireworks/models/llama-v3p1-8b-instruct"]
|
|
1536
1626
|
},
|
|
1537
|
-
"fireworks:mixtral-8x22b-instruct": {
|
|
1538
|
-
id: "fireworks:mixtral-8x22b-instruct",
|
|
1627
|
+
"fireworks-ai:mixtral-8x22b-instruct": {
|
|
1628
|
+
id: "fireworks-ai:mixtral-8x22b-instruct",
|
|
1539
1629
|
name: "Mixtral MoE 8x22B Instruct",
|
|
1540
1630
|
description: "Mistral MoE 8x22B Instruct v0.1 model with Sparse Mixture of Experts. Fine tuned for instruction following.",
|
|
1541
1631
|
input: {
|
|
@@ -1550,8 +1640,8 @@ var models = {
|
|
|
1550
1640
|
lifecycle: "live",
|
|
1551
1641
|
aliases: ["accounts/fireworks/models/mixtral-8x22b-instruct"]
|
|
1552
1642
|
},
|
|
1553
|
-
"fireworks:mixtral-8x7b-instruct": {
|
|
1554
|
-
id: "fireworks:mixtral-8x7b-instruct",
|
|
1643
|
+
"fireworks-ai:mixtral-8x7b-instruct": {
|
|
1644
|
+
id: "fireworks-ai:mixtral-8x7b-instruct",
|
|
1555
1645
|
name: "Mixtral MoE 8x7B Instruct",
|
|
1556
1646
|
description: "Mistral MoE 8x7B Instruct v0.1 model with Sparse Mixture of Experts. Fine tuned for instruction following",
|
|
1557
1647
|
input: {
|
|
@@ -1566,8 +1656,8 @@ var models = {
|
|
|
1566
1656
|
lifecycle: "live",
|
|
1567
1657
|
aliases: ["accounts/fireworks/models/mixtral-8x7b-instruct"]
|
|
1568
1658
|
},
|
|
1569
|
-
"fireworks:mythomax-l2-13b": {
|
|
1570
|
-
id: "fireworks:mythomax-l2-13b",
|
|
1659
|
+
"fireworks-ai:mythomax-l2-13b": {
|
|
1660
|
+
id: "fireworks-ai:mythomax-l2-13b",
|
|
1571
1661
|
name: "MythoMax L2 13b",
|
|
1572
1662
|
description: "MythoMax L2 is designed to excel at both roleplaying and storytelling, and is an improved variant of the previous MythoMix model, combining the MythoLogic-L2 and Huginn models.",
|
|
1573
1663
|
input: {
|
|
@@ -1582,8 +1672,8 @@ var models = {
|
|
|
1582
1672
|
lifecycle: "live",
|
|
1583
1673
|
aliases: ["accounts/fireworks/models/mythomax-l2-13b"]
|
|
1584
1674
|
},
|
|
1585
|
-
"fireworks:gemma2-9b-it": {
|
|
1586
|
-
id: "fireworks:gemma2-9b-it",
|
|
1675
|
+
"fireworks-ai:gemma2-9b-it": {
|
|
1676
|
+
id: "fireworks-ai:gemma2-9b-it",
|
|
1587
1677
|
name: "Gemma 2 9B Instruct",
|
|
1588
1678
|
description: "Redesigned for outsized performance and unmatched efficiency, Gemma 2 optimizes for blazing-fast inference on diverse hardware. Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning.",
|
|
1589
1679
|
input: {
|
|
@@ -1717,7 +1807,7 @@ var CognitiveBeta = class {
|
|
|
1717
1807
|
}
|
|
1718
1808
|
const res = await this._withServerRetry(
|
|
1719
1809
|
() => this._axiosClient.post(
|
|
1720
|
-
"/
|
|
1810
|
+
"/v2/cognitive/generate-text-stream",
|
|
1721
1811
|
{ ...request, stream: true },
|
|
1722
1812
|
{
|
|
1723
1813
|
responseType: "stream",
|