@mastra/core 0.2.0-alpha.83 → 0.2.0-alpha.86

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (116) hide show
  1. package/README.md +1 -1
  2. package/dist/action/index.d.ts +3 -3
  3. package/dist/agent/index.d.ts +6 -6
  4. package/dist/agent/index.js +6 -6
  5. package/dist/base.d.ts +3 -3
  6. package/dist/base.js +2 -2
  7. package/dist/bundler/index.d.ts +21 -0
  8. package/dist/bundler/index.js +4 -0
  9. package/dist/chunk-2ISN3AA7.js +392 -0
  10. package/dist/chunk-2J5OHBUG.js +24 -0
  11. package/dist/{chunk-Y7TKCKRI.js → chunk-5NQ3MEZM.js} +8 -8
  12. package/dist/{chunk-K7KUCZT2.js → chunk-6WJREZ5F.js} +2 -2
  13. package/dist/{chunk-RPBNMIV7.js → chunk-73XDWPXJ.js} +84 -52
  14. package/dist/chunk-CQYUMNLZ.js +28 -0
  15. package/dist/{chunk-AO76Z4K3.js → chunk-D66E7L7R.js} +1 -1
  16. package/dist/{chunk-FCRDS4YS.js → chunk-EH2PG7Q6.js} +10 -4
  17. package/dist/{chunk-BWEOV64I.js → chunk-EVYBUFXB.js} +13 -9
  18. package/dist/{chunk-OJJ6QGX4.js → chunk-I3MJB67Z.js} +12 -12
  19. package/dist/{chunk-27EL4TRB.js → chunk-JCRGAEY6.js} +1 -1
  20. package/dist/{chunk-BQMCLBCR.js → chunk-OOZGW5CY.js} +1 -1
  21. package/dist/{chunk-LYNJHHQX.js → chunk-RCS7AVH6.js} +2 -2
  22. package/dist/chunk-RLPH6TDJ.js +30 -0
  23. package/dist/{chunk-FROFMREE.js → chunk-RZYBOMIG.js} +7 -0
  24. package/dist/{chunk-Q5JF3TKZ.js → chunk-SLWM3SEH.js} +1 -1
  25. package/dist/{chunk-NLQVQJOD.js → chunk-TJK6TGSR.js} +1 -0
  26. package/dist/{chunk-S45F6AHA.js → chunk-U25DBVJZ.js} +1 -1
  27. package/dist/chunk-XD5IQZRB.js +16 -0
  28. package/dist/deployer/index.d.ts +8 -25
  29. package/dist/deployer/index.js +4 -3
  30. package/dist/embeddings/index.d.ts +2 -2
  31. package/dist/embeddings/index.js +1 -1
  32. package/dist/engine/index.d.ts +2 -2
  33. package/dist/engine/index.js +3 -3
  34. package/dist/eval/index.d.ts +3 -3
  35. package/dist/filter/index.d.ts +7 -6
  36. package/dist/filter/index.js +1 -1
  37. package/dist/{index-B9i6dmc0.d.ts → index-B0-NXUYv.d.ts} +1 -0
  38. package/dist/{index-CA9EWqaj.d.ts → index-Duqv1Yom.d.ts} +50 -27
  39. package/dist/index.d.ts +15 -13
  40. package/dist/index.js +26 -25
  41. package/dist/integration/index.d.ts +4 -4
  42. package/dist/llm/index.d.ts +3 -3
  43. package/dist/llm/index.js +3 -3
  44. package/dist/llm/model/providers/anthropic-vertex.d.ts +31 -0
  45. package/dist/llm/model/providers/anthropic-vertex.js +23 -0
  46. package/dist/llm/model/providers/anthropic.d.ts +29 -0
  47. package/dist/llm/model/providers/anthropic.js +21 -0
  48. package/dist/llm/model/providers/azure.d.ts +48 -0
  49. package/dist/llm/model/providers/azure.js +50 -0
  50. package/dist/llm/model/providers/baseten.d.ts +33 -0
  51. package/dist/llm/model/providers/baseten.js +29 -0
  52. package/dist/llm/model/providers/bedrock.d.ts +32 -0
  53. package/dist/llm/model/providers/bedrock.js +24 -0
  54. package/dist/llm/model/providers/cerebras.d.ts +30 -0
  55. package/dist/llm/model/providers/cerebras.js +22 -0
  56. package/dist/llm/model/providers/cohere.d.ts +30 -0
  57. package/dist/llm/model/providers/cohere.js +22 -0
  58. package/dist/llm/model/providers/deepinfra.d.ts +30 -0
  59. package/dist/llm/model/providers/deepinfra.js +22 -0
  60. package/dist/llm/model/providers/deepseek.d.ts +30 -0
  61. package/dist/llm/model/providers/deepseek.js +22 -0
  62. package/dist/llm/model/providers/fireworks.d.ts +35 -0
  63. package/dist/llm/model/providers/fireworks.js +40 -0
  64. package/dist/llm/model/providers/google-vertex.d.ts +48 -0
  65. package/dist/llm/model/providers/google-vertex.js +22 -0
  66. package/dist/llm/model/providers/google.d.ts +54 -0
  67. package/dist/llm/model/providers/google.js +23 -0
  68. package/dist/llm/model/providers/grok.d.ts +32 -0
  69. package/dist/llm/model/providers/grok.js +22 -0
  70. package/dist/llm/model/providers/groq.d.ts +37 -0
  71. package/dist/llm/model/providers/groq.js +42 -0
  72. package/dist/llm/model/providers/lmstudio.d.ts +29 -0
  73. package/dist/llm/model/providers/lmstudio.js +22 -0
  74. package/dist/llm/model/providers/mistral.d.ts +30 -0
  75. package/dist/llm/model/providers/mistral.js +22 -0
  76. package/dist/llm/model/providers/mock.d.ts +30 -0
  77. package/dist/llm/model/providers/mock.js +83 -0
  78. package/dist/llm/model/providers/ollama.d.ts +31 -0
  79. package/dist/llm/model/providers/ollama.js +23 -0
  80. package/dist/llm/model/providers/openai-compat.d.ts +39 -0
  81. package/dist/llm/model/providers/openai-compat.js +6 -0
  82. package/dist/llm/model/providers/openai.d.ts +34 -0
  83. package/dist/llm/model/providers/openai.js +6 -0
  84. package/dist/llm/model/providers/openai.test.d.ts +2 -0
  85. package/dist/llm/model/providers/openai.test.js +220 -0
  86. package/dist/llm/model/providers/perplexity.d.ts +30 -0
  87. package/dist/llm/model/providers/perplexity.js +22 -0
  88. package/dist/llm/model/providers/portkey.d.ts +34 -0
  89. package/dist/llm/model/providers/portkey.js +22 -0
  90. package/dist/llm/model/providers/togetherai.d.ts +30 -0
  91. package/dist/llm/model/providers/togetherai.js +22 -0
  92. package/dist/logger/index.d.ts +1 -1
  93. package/dist/logger/index.js +1 -1
  94. package/dist/mastra/index.d.ts +5 -4
  95. package/dist/mastra/index.js +5 -5
  96. package/dist/memory/index.d.ts +3 -3
  97. package/dist/memory/index.js +3 -3
  98. package/dist/model-QGWIMOSx.d.ts +31 -0
  99. package/dist/relevance/index.d.ts +3 -3
  100. package/dist/relevance/index.js +7 -7
  101. package/dist/storage/index.d.ts +4 -4
  102. package/dist/storage/index.js +3 -3
  103. package/dist/telemetry/index.d.ts +5 -103
  104. package/dist/telemetry/index.js +1 -1
  105. package/dist/telemetry-oCUM52DG.d.ts +104 -0
  106. package/dist/tools/index.d.ts +4 -4
  107. package/dist/tts/index.d.ts +2 -2
  108. package/dist/tts/index.js +4 -4
  109. package/dist/vector/index.d.ts +2 -2
  110. package/dist/vector/index.js +4 -4
  111. package/dist/{workflow-Dam5GtdA.d.ts → workflow-DQ8CtzzU.d.ts} +1 -1
  112. package/dist/workflows/index.d.ts +5 -5
  113. package/dist/workflows/index.js +3 -3
  114. package/package.json +30 -15
  115. package/dist/chunk-5R6DBXQM.js +0 -57
  116. /package/dist/{chunk-OZZS5Y3C.js → chunk-SDKEPBBH.js} +0 -0
@@ -0,0 +1,48 @@
1
+ import { OpenAIChatSettings } from '@ai-sdk/openai/internal';
2
+ import { M as MastraLLM } from '../../../model-QGWIMOSx.js';
3
+ import 'ai';
4
+ import 'json-schema';
5
+ import 'zod';
6
+ import '../../../index-Duqv1Yom.js';
7
+ import '../../../base.js';
8
+ import '@opentelemetry/api';
9
+ import '../../../index-B0-NXUYv.js';
10
+ import 'pino';
11
+ import 'stream';
12
+ import '../../../telemetry-oCUM52DG.js';
13
+ import '@opentelemetry/sdk-node';
14
+ import '@opentelemetry/sdk-trace-base';
15
+ import '../../../metric-D2V4CR8D.js';
16
+ import 'sift';
17
+ import '../../../types-M16hSruO.js';
18
+ import '../../../vector/index.js';
19
+ import '../../../engine-EwEG-4Iv.js';
20
+ import '../../../tts/index.js';
21
+
22
+ type AzureModel = 'gpt-35-turbo-instruct' | (string & {});
23
+ declare class Azure extends MastraLLM {
24
+ constructor({ name, resourceName, apiKey, apiVersion, baseURL, headers, fetch, settings, }: {
25
+ name?: AzureModel;
26
+ resourceName?: string;
27
+ apiKey?: string;
28
+ apiVersion?: string;
29
+ baseURL?: string;
30
+ headers?: Record<string, string>;
31
+ fetch?: typeof globalThis.fetch;
32
+ settings?: OpenAIChatSettings;
33
+ });
34
+ }
35
+ declare class AzureReasoning extends MastraLLM {
36
+ constructor({ name, resourceName, apiKey, apiVersion, baseURL, headers, fetch, settings, }: {
37
+ name?: AzureModel;
38
+ resourceName?: string;
39
+ apiKey?: string;
40
+ apiVersion?: string;
41
+ baseURL?: string;
42
+ headers?: Record<string, string>;
43
+ fetch?: typeof globalThis.fetch;
44
+ settings?: OpenAIChatSettings;
45
+ });
46
+ }
47
+
48
+ export { Azure, type AzureModel, AzureReasoning };
@@ -0,0 +1,50 @@
1
+ import { MastraLLM } from '../../../chunk-2ISN3AA7.js';
2
+ import '../../../chunk-LUULSM4U.js';
3
+ import '../../../chunk-JCRGAEY6.js';
4
+ import '../../../chunk-TJK6TGSR.js';
5
+ import { __name } from '../../../chunk-AJJZUHB4.js';
6
+ import { createAzure } from '@ai-sdk/azure';
7
+ import { wrapLanguageModel, extractReasoningMiddleware } from 'ai';
8
+
9
+ var _Azure = class _Azure extends MastraLLM {
10
+ constructor({ name = "gpt-35-turbo-instruct", resourceName = process.env.AZURE_RESOURCE_NAME || "", apiKey = process.env.AZURE_API_KEY || "", apiVersion, baseURL, headers, fetch, settings }) {
11
+ const azure = createAzure({
12
+ resourceName,
13
+ apiKey,
14
+ apiVersion,
15
+ baseURL,
16
+ headers,
17
+ fetch
18
+ });
19
+ super({
20
+ model: azure(name, settings)
21
+ });
22
+ }
23
+ };
24
+ __name(_Azure, "Azure");
25
+ var Azure = _Azure;
26
+ var _AzureReasoning = class _AzureReasoning extends MastraLLM {
27
+ constructor({ name = "gpt-35-turbo-instruct", resourceName = process.env.AZURE_RESOURCE_NAME || "", apiKey = process.env.AZURE_API_KEY || "", apiVersion, baseURL, headers, fetch, settings }) {
28
+ const azure = createAzure({
29
+ resourceName,
30
+ apiKey,
31
+ apiVersion,
32
+ baseURL,
33
+ headers,
34
+ fetch
35
+ });
36
+ const enhancedModel = wrapLanguageModel({
37
+ model: azure(name, settings),
38
+ middleware: extractReasoningMiddleware({
39
+ tagName: "think"
40
+ })
41
+ });
42
+ super({
43
+ model: enhancedModel
44
+ });
45
+ }
46
+ };
47
+ __name(_AzureReasoning, "AzureReasoning");
48
+ var AzureReasoning = _AzureReasoning;
49
+
50
+ export { Azure, AzureReasoning };
@@ -0,0 +1,33 @@
1
+ import { OpenAIChatSettings } from '@ai-sdk/openai/internal';
2
+ import { M as MastraLLM } from '../../../model-QGWIMOSx.js';
3
+ import 'ai';
4
+ import 'json-schema';
5
+ import 'zod';
6
+ import '../../../index-Duqv1Yom.js';
7
+ import '../../../base.js';
8
+ import '@opentelemetry/api';
9
+ import '../../../index-B0-NXUYv.js';
10
+ import 'pino';
11
+ import 'stream';
12
+ import '../../../telemetry-oCUM52DG.js';
13
+ import '@opentelemetry/sdk-node';
14
+ import '@opentelemetry/sdk-trace-base';
15
+ import '../../../metric-D2V4CR8D.js';
16
+ import 'sift';
17
+ import '../../../types-M16hSruO.js';
18
+ import '../../../vector/index.js';
19
+ import '../../../engine-EwEG-4Iv.js';
20
+ import '../../../tts/index.js';
21
+
22
+ type BasetenModel = 'llama-3.1-70b-instruct' | 'qwen2.5-7b-math-instruct' | 'qwen2.5-14b-instruct' | 'qwen2.5-32b-coder-instruct' | 'llama-3.1-8b-instruct' | 'llama-3.1-nemetron-70b' | 'llama-3.2-90b-vision-instruct' | 'llama-3.1-405b-instruct' | 'ultravox-v0.4' | 'llama-3.2-1b-vision-instruct' | 'llama-3-70b-instruct' | 'llama-3-8b-instruct' | 'mistral-7b-instruct' | 'qwen2.5-14b-coder-instruct' | 'qwen2.5-7b-coder-instruct' | 'qwen2.5-72b-math-instruct' | 'qwen2.5-72b-instruct' | 'qwen2.5-32b-instruct' | 'qwen2.5-7b-instruct' | 'qwen2.5-3b-instruct' | 'pixtral-12b' | 'phi-3.5-mini-instruct' | 'gemma-2-9b' | 'gemma-2-27b' | 'phi-3-mini-128k-instruct' | 'phi-3-mini-4k-instruct' | 'zephyr-7b-alpha' | 'mixtral-8x7b-instruct' | 'mixtral-8x22b' | (string & {});
23
+ declare class BaseTen extends MastraLLM {
24
+ constructor({ name, apiKey, baseURL, fetch, settings, }?: {
25
+ name?: BasetenModel;
26
+ apiKey?: string;
27
+ baseURL?: string;
28
+ fetch?: typeof globalThis.fetch;
29
+ settings?: OpenAIChatSettings;
30
+ });
31
+ }
32
+
33
+ export { BaseTen, type BasetenModel };
@@ -0,0 +1,29 @@
1
+ import { openaiCompat } from '../../../chunk-RLPH6TDJ.js';
2
+ import { MastraLLM } from '../../../chunk-2ISN3AA7.js';
3
+ import '../../../chunk-LUULSM4U.js';
4
+ import '../../../chunk-JCRGAEY6.js';
5
+ import '../../../chunk-TJK6TGSR.js';
6
+ import { __name } from '../../../chunk-AJJZUHB4.js';
7
+
8
+ // src/llm/model/providers/baseten.ts
9
+ var _BaseTen = class _BaseTen extends MastraLLM {
10
+ constructor({ name = "llama-3.1-70b-instruct", apiKey = process.env.BASETEN_API_KEY || "", baseURL = "https://bridge.baseten.co/v1/direct", fetch, settings } = {}) {
11
+ if (fetch) {
12
+ throw new Error("Custom fetch is required to use BaseTen. See https://docs.baseten.co/api-reference/openai for more information");
13
+ }
14
+ const baseten = openaiCompat({
15
+ baseURL,
16
+ modelName: name,
17
+ apiKey,
18
+ fetch,
19
+ settings
20
+ });
21
+ super({
22
+ model: baseten
23
+ });
24
+ }
25
+ };
26
+ __name(_BaseTen, "BaseTen");
27
+ var BaseTen = _BaseTen;
28
+
29
+ export { BaseTen };
@@ -0,0 +1,32 @@
1
+ import { M as MastraLLM } from '../../../model-QGWIMOSx.js';
2
+ import 'ai';
3
+ import 'json-schema';
4
+ import 'zod';
5
+ import '../../../index-Duqv1Yom.js';
6
+ import '../../../base.js';
7
+ import '@opentelemetry/api';
8
+ import '../../../index-B0-NXUYv.js';
9
+ import 'pino';
10
+ import 'stream';
11
+ import '../../../telemetry-oCUM52DG.js';
12
+ import '@opentelemetry/sdk-node';
13
+ import '@opentelemetry/sdk-trace-base';
14
+ import '../../../metric-D2V4CR8D.js';
15
+ import 'sift';
16
+ import '../../../types-M16hSruO.js';
17
+ import '../../../vector/index.js';
18
+ import '../../../engine-EwEG-4Iv.js';
19
+ import '../../../tts/index.js';
20
+
21
+ type AmazonModel = 'amazon-titan-tg1-large' | 'amazon-titan-text-express-v1' | 'anthropic-claude-3-5-sonnet-20241022-v2:0' | 'anthropic-claude-3-5-sonnet-20240620-v1:0' | 'anthropic-claude-3-5-haiku-20241022-v1:0' | 'anthropic-claude-3-opus-20240229-v1:0' | 'anthropic-claude-3-sonnet-20240229-v1:0' | 'anthropic-claude-3-haiku-20240307-v1:0' | 'anthropic-claude-v2:1' | 'cohere-command-r-v1:0' | 'cohere-command-r-plus-v1:0' | 'meta-llama2-13b-chat-v1' | 'meta-llama2-70b-chat-v1' | 'meta-llama3-8b-instruct-v1:0' | 'meta-llama3-70b-instruct-v1:0' | 'meta-llama3-1-8b-instruct-v1:0' | 'meta-llama3-1-70b-instruct-v1:0' | 'meta-llama3-1-405b-instruct-v1:0' | 'meta-llama3-2-1b-instruct-v1:0' | 'meta-llama3-2-3b-instruct-v1:0' | 'meta-llama3-2-11b-instruct-v1:0' | 'meta-llama3-2-90b-instruct-v1:0' | 'mistral-mistral-7b-instruct-v0:2' | 'mistral-mixtral-8x7b-instruct-v0:1' | 'mistral-mistral-large-2402-v1:0' | 'mistral-mistral-small-2402-v1:0' | (string & {});
22
+ declare class AmazonBedrock extends MastraLLM {
23
+ constructor({ name, region, accessKeyId, secretAccessKey, sessionToken, }: {
24
+ name?: AmazonModel;
25
+ region?: string;
26
+ accessKeyId?: string;
27
+ secretAccessKey?: string;
28
+ sessionToken?: string;
29
+ });
30
+ }
31
+
32
+ export { AmazonBedrock, type AmazonModel };
@@ -0,0 +1,24 @@
1
+ import { MastraLLM } from '../../../chunk-2ISN3AA7.js';
2
+ import '../../../chunk-LUULSM4U.js';
3
+ import '../../../chunk-JCRGAEY6.js';
4
+ import '../../../chunk-TJK6TGSR.js';
5
+ import { __name } from '../../../chunk-AJJZUHB4.js';
6
+ import { createAmazonBedrock } from '@ai-sdk/amazon-bedrock';
7
+
8
+ var _AmazonBedrock = class _AmazonBedrock extends MastraLLM {
9
+ constructor({ name = "amazon-titan-tg1-large", region = process.env.AWS_REGION || "", accessKeyId = process.env.AWS_ACCESS_KEY_ID || "", secretAccessKey = process.env.AWS_SECRET_ACCESS_KEY || "", sessionToken = process.env.AWS_SESSION_TOKEN || "" }) {
10
+ const amazon = createAmazonBedrock({
11
+ region,
12
+ accessKeyId,
13
+ secretAccessKey,
14
+ sessionToken
15
+ });
16
+ super({
17
+ model: amazon(name)
18
+ });
19
+ }
20
+ };
21
+ __name(_AmazonBedrock, "AmazonBedrock");
22
+ var AmazonBedrock = _AmazonBedrock;
23
+
24
+ export { AmazonBedrock };
@@ -0,0 +1,30 @@
1
+ import { M as MastraLLM } from '../../../model-QGWIMOSx.js';
2
+ import 'ai';
3
+ import 'json-schema';
4
+ import 'zod';
5
+ import '../../../index-Duqv1Yom.js';
6
+ import '../../../base.js';
7
+ import '@opentelemetry/api';
8
+ import '../../../index-B0-NXUYv.js';
9
+ import 'pino';
10
+ import 'stream';
11
+ import '../../../telemetry-oCUM52DG.js';
12
+ import '@opentelemetry/sdk-node';
13
+ import '@opentelemetry/sdk-trace-base';
14
+ import '../../../metric-D2V4CR8D.js';
15
+ import 'sift';
16
+ import '../../../types-M16hSruO.js';
17
+ import '../../../vector/index.js';
18
+ import '../../../engine-EwEG-4Iv.js';
19
+ import '../../../tts/index.js';
20
+
21
+ type CerebrasModel = 'llama3.1-8b' | 'llama3.1-70b' | 'llama3.3-70b' | (string & {});
22
+ declare class Cerebras extends MastraLLM {
23
+ constructor({ name, apiKey, baseURL, }?: {
24
+ name?: CerebrasModel;
25
+ apiKey?: string;
26
+ baseURL?: string;
27
+ });
28
+ }
29
+
30
+ export { Cerebras, type CerebrasModel };
@@ -0,0 +1,22 @@
1
+ import { MastraLLM } from '../../../chunk-2ISN3AA7.js';
2
+ import '../../../chunk-LUULSM4U.js';
3
+ import '../../../chunk-JCRGAEY6.js';
4
+ import '../../../chunk-TJK6TGSR.js';
5
+ import { __name } from '../../../chunk-AJJZUHB4.js';
6
+ import { createCerebras } from '@ai-sdk/cerebras';
7
+
8
+ var _Cerebras = class _Cerebras extends MastraLLM {
9
+ constructor({ name = "llama3.1-8b", apiKey = process.env.CEREBRAS_API_KEY || "", baseURL = "https://api.cerebras.ai/v1" } = {}) {
10
+ const cerebrasModel = createCerebras({
11
+ baseURL,
12
+ apiKey
13
+ });
14
+ super({
15
+ model: cerebrasModel(name)
16
+ });
17
+ }
18
+ };
19
+ __name(_Cerebras, "Cerebras");
20
+ var Cerebras = _Cerebras;
21
+
22
+ export { Cerebras };
@@ -0,0 +1,30 @@
1
+ import { M as MastraLLM } from '../../../model-QGWIMOSx.js';
2
+ import 'ai';
3
+ import 'json-schema';
4
+ import 'zod';
5
+ import '../../../index-Duqv1Yom.js';
6
+ import '../../../base.js';
7
+ import '@opentelemetry/api';
8
+ import '../../../index-B0-NXUYv.js';
9
+ import 'pino';
10
+ import 'stream';
11
+ import '../../../telemetry-oCUM52DG.js';
12
+ import '@opentelemetry/sdk-node';
13
+ import '@opentelemetry/sdk-trace-base';
14
+ import '../../../metric-D2V4CR8D.js';
15
+ import 'sift';
16
+ import '../../../types-M16hSruO.js';
17
+ import '../../../vector/index.js';
18
+ import '../../../engine-EwEG-4Iv.js';
19
+ import '../../../tts/index.js';
20
+
21
+ type CohereModel = 'command-r-plus' | (string & {});
22
+ declare class Cohere extends MastraLLM {
23
+ constructor({ name, apiKey, baseURL, }: {
24
+ name?: CohereModel;
25
+ apiKey?: string;
26
+ baseURL?: string;
27
+ });
28
+ }
29
+
30
+ export { Cohere, type CohereModel };
@@ -0,0 +1,22 @@
1
+ import { MastraLLM } from '../../../chunk-2ISN3AA7.js';
2
+ import '../../../chunk-LUULSM4U.js';
3
+ import '../../../chunk-JCRGAEY6.js';
4
+ import '../../../chunk-TJK6TGSR.js';
5
+ import { __name } from '../../../chunk-AJJZUHB4.js';
6
+ import { createCohere } from '@ai-sdk/cohere';
7
+
8
+ var _Cohere = class _Cohere extends MastraLLM {
9
+ constructor({ name = "command-r-plus", apiKey = process.env.COHERE_API_KEY || "", baseURL = "https://api.cohere.com/v2" }) {
10
+ const cohere = createCohere({
11
+ baseURL,
12
+ apiKey
13
+ });
14
+ super({
15
+ model: cohere(name)
16
+ });
17
+ }
18
+ };
19
+ __name(_Cohere, "Cohere");
20
+ var Cohere = _Cohere;
21
+
22
+ export { Cohere };
@@ -0,0 +1,30 @@
1
+ import { M as MastraLLM } from '../../../model-QGWIMOSx.js';
2
+ import 'ai';
3
+ import 'json-schema';
4
+ import 'zod';
5
+ import '../../../index-Duqv1Yom.js';
6
+ import '../../../base.js';
7
+ import '@opentelemetry/api';
8
+ import '../../../index-B0-NXUYv.js';
9
+ import 'pino';
10
+ import 'stream';
11
+ import '../../../telemetry-oCUM52DG.js';
12
+ import '@opentelemetry/sdk-node';
13
+ import '@opentelemetry/sdk-trace-base';
14
+ import '../../../metric-D2V4CR8D.js';
15
+ import 'sift';
16
+ import '../../../types-M16hSruO.js';
17
+ import '../../../vector/index.js';
18
+ import '../../../engine-EwEG-4Iv.js';
19
+ import '../../../tts/index.js';
20
+
21
+ type DeepInfraModel = 'meta-llama/Llama-3.3-70B-Instruct-Turbo' | 'meta-llama/Llama-3.3-70B-Instruct' | 'meta-llama/Meta-Llama-3.1-405B-Instruct' | 'meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo' | 'meta-llama/Meta-Llama-3.1-70B-Instruct' | 'meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo' | 'meta-llama/Meta-Llama-3.1-8B-Instruct' | 'meta-llama/Llama-3.2-11B-Vision-Instruct' | 'meta-llama/Llama-3.2-90B-Vision-Instruct' | 'mistralai/Mixtral-8x7B-Instruct-v0.1' | 'deepseek-ai/DeepSeek-V3' | 'nvidia/Llama-3.1-Nemotron-70B-Instruct' | 'Qwen/Qwen2-7B-Instruct' | 'Qwen/Qwen2.5-72B-Instruct' | 'Qwen/Qwen2.5-Coder-32B-Instruct' | 'Qwen/QwQ-32B-Preview' | 'google/codegemma-7b-it' | 'google/gemma-2-9b-it' | 'microsoft/WizardLM-2-8x22B' | (string & {});
22
+ declare class DeepInfra extends MastraLLM {
23
+ constructor({ name, apiKey, baseURL, }?: {
24
+ name?: DeepInfraModel;
25
+ apiKey?: string;
26
+ baseURL?: string;
27
+ });
28
+ }
29
+
30
+ export { DeepInfra, type DeepInfraModel };
@@ -0,0 +1,22 @@
1
+ import { MastraLLM } from '../../../chunk-2ISN3AA7.js';
2
+ import '../../../chunk-LUULSM4U.js';
3
+ import '../../../chunk-JCRGAEY6.js';
4
+ import '../../../chunk-TJK6TGSR.js';
5
+ import { __name } from '../../../chunk-AJJZUHB4.js';
6
+ import { createDeepInfra } from '@ai-sdk/deepinfra';
7
+
8
+ var _DeepInfra = class _DeepInfra extends MastraLLM {
9
+ constructor({ name = "meta-llama/Meta-Llama-3.1-70B-Instruct", apiKey = process.env.DEEPINFRA_API_KEY || "", baseURL = "https://api.deepinfra.com/v1/openai" } = {}) {
10
+ const deepinfraModel = createDeepInfra({
11
+ baseURL,
12
+ apiKey
13
+ });
14
+ super({
15
+ model: deepinfraModel(name)
16
+ });
17
+ }
18
+ };
19
+ __name(_DeepInfra, "DeepInfra");
20
+ var DeepInfra = _DeepInfra;
21
+
22
+ export { DeepInfra };
@@ -0,0 +1,30 @@
1
+ import { M as MastraLLM } from '../../../model-QGWIMOSx.js';
2
+ import 'ai';
3
+ import 'json-schema';
4
+ import 'zod';
5
+ import '../../../index-Duqv1Yom.js';
6
+ import '../../../base.js';
7
+ import '@opentelemetry/api';
8
+ import '../../../index-B0-NXUYv.js';
9
+ import 'pino';
10
+ import 'stream';
11
+ import '../../../telemetry-oCUM52DG.js';
12
+ import '@opentelemetry/sdk-node';
13
+ import '@opentelemetry/sdk-trace-base';
14
+ import '../../../metric-D2V4CR8D.js';
15
+ import 'sift';
16
+ import '../../../types-M16hSruO.js';
17
+ import '../../../vector/index.js';
18
+ import '../../../engine-EwEG-4Iv.js';
19
+ import '../../../tts/index.js';
20
+
21
+ type DeepseekModel = 'deepseek-chat' | 'deepseek-reasoner' | (string & {});
22
+ declare class DeepSeek extends MastraLLM {
23
+ constructor({ name, apiKey, baseURL, }?: {
24
+ name?: string;
25
+ apiKey?: string;
26
+ baseURL?: string;
27
+ });
28
+ }
29
+
30
+ export { DeepSeek, type DeepseekModel };
@@ -0,0 +1,22 @@
1
+ import { MastraLLM } from '../../../chunk-2ISN3AA7.js';
2
+ import '../../../chunk-LUULSM4U.js';
3
+ import '../../../chunk-JCRGAEY6.js';
4
+ import '../../../chunk-TJK6TGSR.js';
5
+ import { __name } from '../../../chunk-AJJZUHB4.js';
6
+ import { createDeepSeek } from '@ai-sdk/deepseek';
7
+
8
+ var _DeepSeek = class _DeepSeek extends MastraLLM {
9
+ constructor({ name = "deepseek-chat", apiKey = process.env.DEEPSEEK_API_KEY || "", baseURL = "https://api.deepseek.com/v1" } = {}) {
10
+ const deepseekModel = createDeepSeek({
11
+ baseURL,
12
+ apiKey
13
+ });
14
+ super({
15
+ model: deepseekModel(name)
16
+ });
17
+ }
18
+ };
19
+ __name(_DeepSeek, "DeepSeek");
20
+ var DeepSeek = _DeepSeek;
21
+
22
+ export { DeepSeek };
@@ -0,0 +1,35 @@
1
+ import { M as MastraLLM } from '../../../model-QGWIMOSx.js';
2
+ import 'ai';
3
+ import 'json-schema';
4
+ import 'zod';
5
+ import '../../../index-Duqv1Yom.js';
6
+ import '../../../base.js';
7
+ import '@opentelemetry/api';
8
+ import '../../../index-B0-NXUYv.js';
9
+ import 'pino';
10
+ import 'stream';
11
+ import '../../../telemetry-oCUM52DG.js';
12
+ import '@opentelemetry/sdk-node';
13
+ import '@opentelemetry/sdk-trace-base';
14
+ import '../../../metric-D2V4CR8D.js';
15
+ import 'sift';
16
+ import '../../../types-M16hSruO.js';
17
+ import '../../../vector/index.js';
18
+ import '../../../engine-EwEG-4Iv.js';
19
+ import '../../../tts/index.js';
20
+
21
+ type FireworksModel = 'accounts/fireworks/models/deepseek-v3' | 'accounts/fireworks/models/llama-v3p1-405b-instruct' | 'accounts/fireworks/models/llama-v3p1-8b-instruct' | 'accounts/fireworks/models/llama-v3p2-3b-instruct' | 'accounts/fireworks/models/llama-v3p3-70b-instruct' | 'accounts/fireworks/models/mixtral-8x7b-instruct-hf' | 'accounts/fireworks/models/mixtral-8x22b-instruct' | 'accounts/fireworks/models/qwen2p5-coder-32b-instruct' | 'accounts/fireworks/models/llama-v3p2-11b-vision-instruct' | 'accounts/fireworks/models/yi-large' | (string & {});
22
+ declare class Fireworks extends MastraLLM {
23
+ constructor({ name, apiKey }: {
24
+ name: FireworksModel;
25
+ apiKey?: string;
26
+ });
27
+ }
28
+ declare class FireworksReasoning extends MastraLLM {
29
+ constructor({ name, apiKey, }: {
30
+ name: 'accounts/fireworks/models/deepseek-r1';
31
+ apiKey?: string;
32
+ });
33
+ }
34
+
35
+ export { Fireworks, type FireworksModel, FireworksReasoning };
@@ -0,0 +1,40 @@
1
+ import { MastraLLM } from '../../../chunk-2ISN3AA7.js';
2
+ import '../../../chunk-LUULSM4U.js';
3
+ import '../../../chunk-JCRGAEY6.js';
4
+ import '../../../chunk-TJK6TGSR.js';
5
+ import { __name } from '../../../chunk-AJJZUHB4.js';
6
+ import { createFireworks } from '@ai-sdk/fireworks';
7
+ import { wrapLanguageModel, extractReasoningMiddleware } from 'ai';
8
+
9
+ var _Fireworks = class _Fireworks extends MastraLLM {
10
+ constructor({ name, apiKey = process.env.FIREWORKS_API_KEY || "" }) {
11
+ const fireworksModel = createFireworks({
12
+ apiKey
13
+ });
14
+ super({
15
+ model: fireworksModel(name)
16
+ });
17
+ }
18
+ };
19
+ __name(_Fireworks, "Fireworks");
20
+ var Fireworks = _Fireworks;
21
+ var _FireworksReasoning = class _FireworksReasoning extends MastraLLM {
22
+ constructor({ name, apiKey = process.env.FIREWORKS_API_KEY || "" }) {
23
+ const fireworksModel = createFireworks({
24
+ apiKey
25
+ });
26
+ const enhancedModel = wrapLanguageModel({
27
+ model: fireworksModel(name),
28
+ middleware: extractReasoningMiddleware({
29
+ tagName: "think"
30
+ })
31
+ });
32
+ super({
33
+ model: enhancedModel
34
+ });
35
+ }
36
+ };
37
+ __name(_FireworksReasoning, "FireworksReasoning");
38
+ var FireworksReasoning = _FireworksReasoning;
39
+
40
+ export { Fireworks, FireworksReasoning };
@@ -0,0 +1,48 @@
1
+ import { M as MastraLLM } from '../../../model-QGWIMOSx.js';
2
+ import 'ai';
3
+ import 'json-schema';
4
+ import 'zod';
5
+ import '../../../index-Duqv1Yom.js';
6
+ import '../../../base.js';
7
+ import '@opentelemetry/api';
8
+ import '../../../index-B0-NXUYv.js';
9
+ import 'pino';
10
+ import 'stream';
11
+ import '../../../telemetry-oCUM52DG.js';
12
+ import '@opentelemetry/sdk-node';
13
+ import '@opentelemetry/sdk-trace-base';
14
+ import '../../../metric-D2V4CR8D.js';
15
+ import 'sift';
16
+ import '../../../types-M16hSruO.js';
17
+ import '../../../vector/index.js';
18
+ import '../../../engine-EwEG-4Iv.js';
19
+ import '../../../tts/index.js';
20
+
21
+ interface GoogleVertexSettings {
22
+ /**
23
+ Optional. The maximum number of tokens to consider when sampling.
24
+
25
+ Models use nucleus sampling or combined Top-k and nucleus sampling.
26
+ Top-k sampling considers the set of topK most probable tokens.
27
+ Models running with nucleus sampling don't allow topK setting.
28
+ */
29
+ topK?: number;
30
+ /**
31
+ Optional. A list of unique safety settings for blocking unsafe content.
32
+ */
33
+ safetySettings?: Array<{
34
+ category: 'HARM_CATEGORY_UNSPECIFIED' | 'HARM_CATEGORY_HATE_SPEECH' | 'HARM_CATEGORY_DANGEROUS_CONTENT' | 'HARM_CATEGORY_HARASSMENT' | 'HARM_CATEGORY_SEXUALLY_EXPLICIT';
35
+ threshold: 'HARM_BLOCK_THRESHOLD_UNSPECIFIED' | 'BLOCK_LOW_AND_ABOVE' | 'BLOCK_MEDIUM_AND_ABOVE' | 'BLOCK_ONLY_HIGH' | 'BLOCK_NONE';
36
+ }>;
37
+ }
38
+ type GoogleVertexModel = 'gemini-2.0-flash-exp' | 'gemini-1.5-flash' | 'gemini-1.5-pro';
39
+ declare class GoogleVertex extends MastraLLM {
40
+ constructor({ name, project, location, settings, }?: {
41
+ name?: GoogleVertexModel;
42
+ project?: string;
43
+ location?: string;
44
+ settings?: GoogleVertexSettings;
45
+ });
46
+ }
47
+
48
+ export { GoogleVertex, type GoogleVertexModel, type GoogleVertexSettings };
@@ -0,0 +1,22 @@
1
+ import { MastraLLM } from '../../../chunk-2ISN3AA7.js';
2
+ import '../../../chunk-LUULSM4U.js';
3
+ import '../../../chunk-JCRGAEY6.js';
4
+ import '../../../chunk-TJK6TGSR.js';
5
+ import { __name } from '../../../chunk-AJJZUHB4.js';
6
+ import { createVertex } from '@ai-sdk/google-vertex';
7
+
8
+ var _GoogleVertex = class _GoogleVertex extends MastraLLM {
9
+ constructor({ name = "gemini-1.5-pro", project = process.env.GOOGLE_VERTEX_PROJECT || "", location = process.env.GOOGLE_VERTEX_LOCATION || "us-central1", settings } = {}) {
10
+ const vertexModel = createVertex({
11
+ project,
12
+ location
13
+ });
14
+ super({
15
+ model: vertexModel(name, settings)
16
+ });
17
+ }
18
+ };
19
+ __name(_GoogleVertex, "GoogleVertex");
20
+ var GoogleVertex = _GoogleVertex;
21
+
22
+ export { GoogleVertex };
@@ -0,0 +1,54 @@
1
+ import { M as MastraLLM } from '../../../model-QGWIMOSx.js';
2
+ import 'ai';
3
+ import 'json-schema';
4
+ import 'zod';
5
+ import '../../../index-Duqv1Yom.js';
6
+ import '../../../base.js';
7
+ import '@opentelemetry/api';
8
+ import '../../../index-B0-NXUYv.js';
9
+ import 'pino';
10
+ import 'stream';
11
+ import '../../../telemetry-oCUM52DG.js';
12
+ import '@opentelemetry/sdk-node';
13
+ import '@opentelemetry/sdk-trace-base';
14
+ import '../../../metric-D2V4CR8D.js';
15
+ import 'sift';
16
+ import '../../../types-M16hSruO.js';
17
+ import '../../../vector/index.js';
18
+ import '../../../engine-EwEG-4Iv.js';
19
+ import '../../../tts/index.js';
20
+
21
+ interface GoogleGenerativeAISettings {
22
+ /**
23
+ Optional.
24
+ The name of the cached content used as context to serve the prediction.
25
+ Format: cachedContents/{cachedContent}
26
+ */
27
+ cachedContent?: string;
28
+ /**
29
+ * Optional. Enable structured output. Default is true.
30
+ *
31
+ * This is useful when the JSON Schema contains elements that are
32
+ * not supported by the OpenAPI schema version that
33
+ * Google Generative AI uses. You can use this to disable
34
+ * structured outputs if you need to.
35
+ */
36
+ structuredOutputs?: boolean;
37
+ /**
38
+ Optional. A list of unique safety settings for blocking unsafe content.
39
+ */
40
+ safetySettings?: Array<{
41
+ category: 'HARM_CATEGORY_HATE_SPEECH' | 'HARM_CATEGORY_DANGEROUS_CONTENT' | 'HARM_CATEGORY_HARASSMENT' | 'HARM_CATEGORY_SEXUALLY_EXPLICIT';
42
+ threshold: 'HARM_BLOCK_THRESHOLD_UNSPECIFIED' | 'BLOCK_LOW_AND_ABOVE' | 'BLOCK_MEDIUM_AND_ABOVE' | 'BLOCK_ONLY_HIGH' | 'BLOCK_NONE';
43
+ }>;
44
+ }
45
+ type GoogleModel = 'gemini-1.5-pro-latest' | 'gemini-1.5-pro' | 'gemini-1.5-flash-latest' | 'gemini-1.5-flash' | 'gemini-2.0-flash-exp-latest' | 'gemini-2.0-flash-thinking-exp-1219' | 'gemini-exp-1206' | (string & {});
46
+ declare class Gemini extends MastraLLM {
47
+ constructor({ name, apiKey, settings, }?: {
48
+ name?: GoogleModel;
49
+ apiKey?: string;
50
+ settings?: GoogleGenerativeAISettings;
51
+ });
52
+ }
53
+
54
+ export { Gemini, type GoogleGenerativeAISettings, type GoogleModel };