@mastra/core 0.2.0-alpha.93 → 0.2.0-alpha.95

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (137) hide show
  1. package/README.md +11 -8
  2. package/dist/action/index.d.ts +1 -3
  3. package/dist/agent/index.d.ts +2 -4
  4. package/dist/agent/index.js +3 -5
  5. package/dist/{chunk-SDKEPBBH.js → chunk-4ZUSEHLH.js} +109 -12
  6. package/dist/{chunk-F7ILHRX5.js → chunk-5DZIXJRV.js} +1 -1
  7. package/dist/chunk-AE3H2QEY.js +1 -0
  8. package/dist/{chunk-IE37CBXB.js → chunk-G4LP2IJU.js} +1 -3
  9. package/dist/{chunk-JWVCENG2.js → chunk-JY322VLW.js} +445 -32
  10. package/dist/{chunk-ECUVL2G5.js → chunk-K3N7KJHH.js} +1 -1
  11. package/dist/chunk-MBOUQZQT.js +16 -0
  12. package/dist/{chunk-GGYXCZUW.js → chunk-N44BCOWP.js} +4 -12
  13. package/dist/{chunk-WBPDZBUT.js → chunk-TYIBRZOY.js} +2 -17
  14. package/dist/eval/index.d.ts +1 -3
  15. package/dist/index-CBxGSZZE.d.ts +808 -0
  16. package/dist/index.d.ts +6 -12
  17. package/dist/index.js +12 -26
  18. package/dist/integration/index.d.ts +5 -7
  19. package/dist/llm/index.d.ts +4 -6
  20. package/dist/llm/index.js +1 -6
  21. package/dist/mastra/index.d.ts +2 -5
  22. package/dist/mastra/index.js +2 -6
  23. package/dist/memory/index.d.ts +1 -3
  24. package/dist/memory/index.js +1 -1
  25. package/dist/relevance/index.d.ts +2 -19
  26. package/dist/relevance/index.js +4 -6
  27. package/dist/storage/index.d.ts +2 -4
  28. package/dist/storage/index.js +2 -3
  29. package/dist/telemetry/index.js +1 -2
  30. package/dist/tools/index.d.ts +2 -4
  31. package/dist/tts/index.js +2 -3
  32. package/dist/vector/index.d.ts +0 -7
  33. package/dist/vector/index.js +1 -2
  34. package/dist/vector/libsql/index.d.ts +0 -1
  35. package/dist/vector/libsql/index.js +2 -3
  36. package/dist/{workflow-CSwxE-4q.d.ts → workflow-BA2Pnq90.d.ts} +1 -1
  37. package/dist/workflows/index.d.ts +3 -5
  38. package/package.json +20 -56
  39. package/dist/chunk-5DYJZVB7.js +0 -66
  40. package/dist/chunk-6ZVFVYLE.js +0 -101
  41. package/dist/chunk-AD6IIOCI.js +0 -11
  42. package/dist/chunk-BOS3IA23.js +0 -76
  43. package/dist/chunk-CQU72KBX.js +0 -54
  44. package/dist/chunk-HH5JIATB.js +0 -36
  45. package/dist/chunk-IEFQZ3IL.js +0 -78
  46. package/dist/chunk-IJ55HGH4.js +0 -24
  47. package/dist/chunk-NNNAWW2H.js +0 -66
  48. package/dist/chunk-NYBGZL6Z.js +0 -30
  49. package/dist/chunk-QX2CHXA7.js +0 -408
  50. package/dist/chunk-RAR7V3E3.js +0 -66
  51. package/dist/chunk-RI3ECMVF.js +0 -636
  52. package/dist/chunk-VPSYTVWP.js +0 -66
  53. package/dist/chunk-ZA7MIXNW.js +0 -66
  54. package/dist/embeddings/index.d.ts +0 -8
  55. package/dist/embeddings/index.js +0 -2
  56. package/dist/embeddings/model/providers/bedrock.d.ts +0 -42
  57. package/dist/embeddings/model/providers/bedrock.js +0 -3
  58. package/dist/embeddings/model/providers/bedrock.test.d.ts +0 -2
  59. package/dist/embeddings/model/providers/bedrock.test.js +0 -37
  60. package/dist/embeddings/model/providers/cohere.d.ts +0 -34
  61. package/dist/embeddings/model/providers/cohere.js +0 -3
  62. package/dist/embeddings/model/providers/cohere.test.d.ts +0 -2
  63. package/dist/embeddings/model/providers/cohere.test.js +0 -37
  64. package/dist/embeddings/model/providers/embedder.d.ts +0 -28
  65. package/dist/embeddings/model/providers/embedder.js +0 -2
  66. package/dist/embeddings/model/providers/google.d.ts +0 -34
  67. package/dist/embeddings/model/providers/google.js +0 -3
  68. package/dist/embeddings/model/providers/google.test.d.ts +0 -2
  69. package/dist/embeddings/model/providers/google.test.js +0 -37
  70. package/dist/embeddings/model/providers/mistral.d.ts +0 -34
  71. package/dist/embeddings/model/providers/mistral.js +0 -3
  72. package/dist/embeddings/model/providers/mistral.test.d.ts +0 -2
  73. package/dist/embeddings/model/providers/mistral.test.js +0 -37
  74. package/dist/embeddings/model/providers/mock.d.ts +0 -20
  75. package/dist/embeddings/model/providers/mock.js +0 -3
  76. package/dist/embeddings/model/providers/mock.test.d.ts +0 -2
  77. package/dist/embeddings/model/providers/mock.test.js +0 -29
  78. package/dist/embeddings/model/providers/openai.d.ts +0 -34
  79. package/dist/embeddings/model/providers/openai.js +0 -3
  80. package/dist/embeddings/model/providers/openai.test.d.ts +0 -2
  81. package/dist/embeddings/model/providers/openai.test.js +0 -37
  82. package/dist/embeddings/model/providers/voyage.d.ts +0 -34
  83. package/dist/embeddings/model/providers/voyage.js +0 -3
  84. package/dist/embeddings/model/providers/voyage.test.d.ts +0 -2
  85. package/dist/embeddings/model/providers/voyage.test.js +0 -37
  86. package/dist/index-B48181D5.d.ts +0 -1070
  87. package/dist/llm/model/index.d.ts +0 -45
  88. package/dist/llm/model/index.js +0 -5
  89. package/dist/llm/model/providers/anthropic-vertex.d.ts +0 -31
  90. package/dist/llm/model/providers/anthropic-vertex.js +0 -23
  91. package/dist/llm/model/providers/anthropic.d.ts +0 -29
  92. package/dist/llm/model/providers/anthropic.js +0 -21
  93. package/dist/llm/model/providers/azure.d.ts +0 -48
  94. package/dist/llm/model/providers/azure.js +0 -50
  95. package/dist/llm/model/providers/baseten.d.ts +0 -33
  96. package/dist/llm/model/providers/baseten.js +0 -29
  97. package/dist/llm/model/providers/bedrock.d.ts +0 -32
  98. package/dist/llm/model/providers/bedrock.js +0 -24
  99. package/dist/llm/model/providers/cerebras.d.ts +0 -30
  100. package/dist/llm/model/providers/cerebras.js +0 -22
  101. package/dist/llm/model/providers/cohere.d.ts +0 -30
  102. package/dist/llm/model/providers/cohere.js +0 -22
  103. package/dist/llm/model/providers/deepinfra.d.ts +0 -30
  104. package/dist/llm/model/providers/deepinfra.js +0 -22
  105. package/dist/llm/model/providers/deepseek.d.ts +0 -30
  106. package/dist/llm/model/providers/deepseek.js +0 -22
  107. package/dist/llm/model/providers/fireworks.d.ts +0 -35
  108. package/dist/llm/model/providers/fireworks.js +0 -40
  109. package/dist/llm/model/providers/google-vertex.d.ts +0 -48
  110. package/dist/llm/model/providers/google-vertex.js +0 -22
  111. package/dist/llm/model/providers/google.d.ts +0 -54
  112. package/dist/llm/model/providers/google.js +0 -23
  113. package/dist/llm/model/providers/grok.d.ts +0 -32
  114. package/dist/llm/model/providers/grok.js +0 -22
  115. package/dist/llm/model/providers/groq.d.ts +0 -37
  116. package/dist/llm/model/providers/groq.js +0 -42
  117. package/dist/llm/model/providers/lmstudio.d.ts +0 -29
  118. package/dist/llm/model/providers/lmstudio.js +0 -22
  119. package/dist/llm/model/providers/mistral.d.ts +0 -30
  120. package/dist/llm/model/providers/mistral.js +0 -22
  121. package/dist/llm/model/providers/mock.d.ts +0 -30
  122. package/dist/llm/model/providers/mock.js +0 -83
  123. package/dist/llm/model/providers/ollama.d.ts +0 -31
  124. package/dist/llm/model/providers/ollama.js +0 -23
  125. package/dist/llm/model/providers/openai-compat.d.ts +0 -39
  126. package/dist/llm/model/providers/openai-compat.js +0 -6
  127. package/dist/llm/model/providers/openai.d.ts +0 -34
  128. package/dist/llm/model/providers/openai.js +0 -6
  129. package/dist/llm/model/providers/openai.test.d.ts +0 -2
  130. package/dist/llm/model/providers/openai.test.js +0 -220
  131. package/dist/llm/model/providers/perplexity.d.ts +0 -30
  132. package/dist/llm/model/providers/perplexity.js +0 -22
  133. package/dist/llm/model/providers/portkey.d.ts +0 -34
  134. package/dist/llm/model/providers/portkey.js +0 -22
  135. package/dist/llm/model/providers/togetherai.d.ts +0 -30
  136. package/dist/llm/model/providers/togetherai.js +0 -22
  137. package/dist/types-M16hSruO.d.ts +0 -40
@@ -1,40 +0,0 @@
1
- import { MastraLLM } from '../../../chunk-QX2CHXA7.js';
2
- import '../../../chunk-KNPBNSJ7.js';
3
- import '../../../chunk-G4MCO7XF.js';
4
- import '../../../chunk-ICMEXHKD.js';
5
- import { __name } from '../../../chunk-AJJZUHB4.js';
6
- import { createFireworks } from '@ai-sdk/fireworks';
7
- import { wrapLanguageModel, extractReasoningMiddleware } from 'ai';
8
-
9
- var _Fireworks = class _Fireworks extends MastraLLM {
10
- constructor({ name, apiKey = process.env.FIREWORKS_API_KEY || "" }) {
11
- const fireworksModel = createFireworks({
12
- apiKey
13
- });
14
- super({
15
- model: fireworksModel(name)
16
- });
17
- }
18
- };
19
- __name(_Fireworks, "Fireworks");
20
- var Fireworks = _Fireworks;
21
- var _FireworksReasoning = class _FireworksReasoning extends MastraLLM {
22
- constructor({ name, apiKey = process.env.FIREWORKS_API_KEY || "" }) {
23
- const fireworksModel = createFireworks({
24
- apiKey
25
- });
26
- const enhancedModel = wrapLanguageModel({
27
- model: fireworksModel(name),
28
- middleware: extractReasoningMiddleware({
29
- tagName: "think"
30
- })
31
- });
32
- super({
33
- model: enhancedModel
34
- });
35
- }
36
- };
37
- __name(_FireworksReasoning, "FireworksReasoning");
38
- var FireworksReasoning = _FireworksReasoning;
39
-
40
- export { Fireworks, FireworksReasoning };
@@ -1,48 +0,0 @@
1
- import { MastraLLM } from '../index.js';
2
- import '../../../index-B48181D5.js';
3
- import 'ai';
4
- import 'json-schema';
5
- import 'zod';
6
- import '../../../base.js';
7
- import '@opentelemetry/api';
8
- import '../../../index-CBZ2mk2H.js';
9
- import 'pino';
10
- import 'stream';
11
- import '../../../telemetry-oCUM52DG.js';
12
- import '@opentelemetry/sdk-node';
13
- import '@opentelemetry/sdk-trace-base';
14
- import '../../../metric-D2V4CR8D.js';
15
- import 'sift';
16
- import '../../../embeddings/model/providers/embedder.js';
17
- import '../../../vector/index.js';
18
- import '../../../types-M16hSruO.js';
19
- import '../../../tts/index.js';
20
-
21
- interface GoogleVertexSettings {
22
- /**
23
- Optional. The maximum number of tokens to consider when sampling.
24
-
25
- Models use nucleus sampling or combined Top-k and nucleus sampling.
26
- Top-k sampling considers the set of topK most probable tokens.
27
- Models running with nucleus sampling don't allow topK setting.
28
- */
29
- topK?: number;
30
- /**
31
- Optional. A list of unique safety settings for blocking unsafe content.
32
- */
33
- safetySettings?: Array<{
34
- category: 'HARM_CATEGORY_UNSPECIFIED' | 'HARM_CATEGORY_HATE_SPEECH' | 'HARM_CATEGORY_DANGEROUS_CONTENT' | 'HARM_CATEGORY_HARASSMENT' | 'HARM_CATEGORY_SEXUALLY_EXPLICIT';
35
- threshold: 'HARM_BLOCK_THRESHOLD_UNSPECIFIED' | 'BLOCK_LOW_AND_ABOVE' | 'BLOCK_MEDIUM_AND_ABOVE' | 'BLOCK_ONLY_HIGH' | 'BLOCK_NONE';
36
- }>;
37
- }
38
- type GoogleVertexModel = 'gemini-2.0-flash-exp' | 'gemini-1.5-flash' | 'gemini-1.5-pro';
39
- declare class GoogleVertex extends MastraLLM {
40
- constructor({ name, project, location, settings, }?: {
41
- name?: GoogleVertexModel;
42
- project?: string;
43
- location?: string;
44
- settings?: GoogleVertexSettings;
45
- });
46
- }
47
-
48
- export { GoogleVertex, type GoogleVertexModel, type GoogleVertexSettings };
@@ -1,22 +0,0 @@
1
- import { MastraLLM } from '../../../chunk-QX2CHXA7.js';
2
- import '../../../chunk-KNPBNSJ7.js';
3
- import '../../../chunk-G4MCO7XF.js';
4
- import '../../../chunk-ICMEXHKD.js';
5
- import { __name } from '../../../chunk-AJJZUHB4.js';
6
- import { createVertex } from '@ai-sdk/google-vertex';
7
-
8
- var _GoogleVertex = class _GoogleVertex extends MastraLLM {
9
- constructor({ name = "gemini-1.5-pro", project = process.env.GOOGLE_VERTEX_PROJECT || "", location = process.env.GOOGLE_VERTEX_LOCATION || "us-central1", settings } = {}) {
10
- const vertexModel = createVertex({
11
- project,
12
- location
13
- });
14
- super({
15
- model: vertexModel(name, settings)
16
- });
17
- }
18
- };
19
- __name(_GoogleVertex, "GoogleVertex");
20
- var GoogleVertex = _GoogleVertex;
21
-
22
- export { GoogleVertex };
@@ -1,54 +0,0 @@
1
- import { MastraLLM } from '../index.js';
2
- import '../../../index-B48181D5.js';
3
- import 'ai';
4
- import 'json-schema';
5
- import 'zod';
6
- import '../../../base.js';
7
- import '@opentelemetry/api';
8
- import '../../../index-CBZ2mk2H.js';
9
- import 'pino';
10
- import 'stream';
11
- import '../../../telemetry-oCUM52DG.js';
12
- import '@opentelemetry/sdk-node';
13
- import '@opentelemetry/sdk-trace-base';
14
- import '../../../metric-D2V4CR8D.js';
15
- import 'sift';
16
- import '../../../embeddings/model/providers/embedder.js';
17
- import '../../../vector/index.js';
18
- import '../../../types-M16hSruO.js';
19
- import '../../../tts/index.js';
20
-
21
- interface GoogleGenerativeAISettings {
22
- /**
23
- Optional.
24
- The name of the cached content used as context to serve the prediction.
25
- Format: cachedContents/{cachedContent}
26
- */
27
- cachedContent?: string;
28
- /**
29
- * Optional. Enable structured output. Default is true.
30
- *
31
- * This is useful when the JSON Schema contains elements that are
32
- * not supported by the OpenAPI schema version that
33
- * Google Generative AI uses. You can use this to disable
34
- * structured outputs if you need to.
35
- */
36
- structuredOutputs?: boolean;
37
- /**
38
- Optional. A list of unique safety settings for blocking unsafe content.
39
- */
40
- safetySettings?: Array<{
41
- category: 'HARM_CATEGORY_HATE_SPEECH' | 'HARM_CATEGORY_DANGEROUS_CONTENT' | 'HARM_CATEGORY_HARASSMENT' | 'HARM_CATEGORY_SEXUALLY_EXPLICIT';
42
- threshold: 'HARM_BLOCK_THRESHOLD_UNSPECIFIED' | 'BLOCK_LOW_AND_ABOVE' | 'BLOCK_MEDIUM_AND_ABOVE' | 'BLOCK_ONLY_HIGH' | 'BLOCK_NONE';
43
- }>;
44
- }
45
- type GoogleModel = 'gemini-1.5-pro-latest' | 'gemini-1.5-pro' | 'gemini-1.5-flash-latest' | 'gemini-1.5-flash' | 'gemini-2.0-flash-exp-latest' | 'gemini-2.0-flash-thinking-exp-1219' | 'gemini-exp-1206' | (string & {});
46
- declare class Gemini extends MastraLLM {
47
- constructor({ name, apiKey, settings, }?: {
48
- name?: GoogleModel;
49
- apiKey?: string;
50
- settings?: GoogleGenerativeAISettings;
51
- });
52
- }
53
-
54
- export { Gemini, type GoogleGenerativeAISettings, type GoogleModel };
@@ -1,23 +0,0 @@
1
- import { MastraLLM } from '../../../chunk-QX2CHXA7.js';
2
- import '../../../chunk-KNPBNSJ7.js';
3
- import '../../../chunk-G4MCO7XF.js';
4
- import '../../../chunk-ICMEXHKD.js';
5
- import { __name } from '../../../chunk-AJJZUHB4.js';
6
- import { createGoogleGenerativeAI } from '@ai-sdk/google';
7
-
8
- var _Gemini = class _Gemini extends MastraLLM {
9
- constructor({ name = "gemini-1.5-pro-latest", apiKey = process.env.GOOGLE_GENERATIVE_AI_API_KEY || "", settings } = {}) {
10
- const google = createGoogleGenerativeAI({
11
- baseURL: "https://generativelanguage.googleapis.com/v1beta",
12
- apiKey
13
- });
14
- const gemini = google(name, settings);
15
- super({
16
- model: gemini
17
- });
18
- }
19
- };
20
- __name(_Gemini, "Gemini");
21
- var Gemini = _Gemini;
22
-
23
- export { Gemini };
@@ -1,32 +0,0 @@
1
- import { OpenAIChatSettings } from '@ai-sdk/openai/internal';
2
- import { MastraLLM } from '../index.js';
3
- import '../../../index-B48181D5.js';
4
- import 'ai';
5
- import 'json-schema';
6
- import 'zod';
7
- import '../../../base.js';
8
- import '@opentelemetry/api';
9
- import '../../../index-CBZ2mk2H.js';
10
- import 'pino';
11
- import 'stream';
12
- import '../../../telemetry-oCUM52DG.js';
13
- import '@opentelemetry/sdk-node';
14
- import '@opentelemetry/sdk-trace-base';
15
- import '../../../metric-D2V4CR8D.js';
16
- import 'sift';
17
- import '../../../embeddings/model/providers/embedder.js';
18
- import '../../../vector/index.js';
19
- import '../../../types-M16hSruO.js';
20
- import '../../../tts/index.js';
21
-
22
- type XGrokModel = 'grok-beta' | 'grok-vision-beta' | 'grok-2-1212' | 'grok-2-vision-1212' | (string & {});
23
- declare class Grok extends MastraLLM {
24
- constructor({ name, apiKey, baseURL, settings, }: {
25
- settings?: OpenAIChatSettings;
26
- name?: string;
27
- apiKey?: string;
28
- baseURL?: string;
29
- });
30
- }
31
-
32
- export { Grok, type XGrokModel };
@@ -1,22 +0,0 @@
1
- import { MastraLLM } from '../../../chunk-QX2CHXA7.js';
2
- import '../../../chunk-KNPBNSJ7.js';
3
- import '../../../chunk-G4MCO7XF.js';
4
- import '../../../chunk-ICMEXHKD.js';
5
- import { __name } from '../../../chunk-AJJZUHB4.js';
6
- import { createXai } from '@ai-sdk/xai';
7
-
8
- var _Grok = class _Grok extends MastraLLM {
9
- constructor({ name = "grok-beta", apiKey = process.env.XAI_API_KEY ?? "", baseURL = "https://api.x.ai/v1", settings }) {
10
- const xAi = createXai({
11
- baseURL,
12
- apiKey
13
- });
14
- super({
15
- model: xAi(name, settings)
16
- });
17
- }
18
- };
19
- __name(_Grok, "Grok");
20
- var Grok = _Grok;
21
-
22
- export { Grok };
@@ -1,37 +0,0 @@
1
- import { MastraLLM } from '../index.js';
2
- import '../../../index-B48181D5.js';
3
- import 'ai';
4
- import 'json-schema';
5
- import 'zod';
6
- import '../../../base.js';
7
- import '@opentelemetry/api';
8
- import '../../../index-CBZ2mk2H.js';
9
- import 'pino';
10
- import 'stream';
11
- import '../../../telemetry-oCUM52DG.js';
12
- import '@opentelemetry/sdk-node';
13
- import '@opentelemetry/sdk-trace-base';
14
- import '../../../metric-D2V4CR8D.js';
15
- import 'sift';
16
- import '../../../embeddings/model/providers/embedder.js';
17
- import '../../../vector/index.js';
18
- import '../../../types-M16hSruO.js';
19
- import '../../../tts/index.js';
20
-
21
- type GroqModel = 'llama3-groq-70b-8192-tool-use-preview' | 'llama3-groq-8b-8192-tool-use-preview' | 'gemma2-9b-it' | 'gemma-7b-it' | (string & {});
22
- declare class Groq extends MastraLLM {
23
- constructor({ name, apiKey, baseURL, }?: {
24
- name?: GroqModel;
25
- apiKey?: string;
26
- baseURL?: string;
27
- });
28
- }
29
- declare class GroqReasoning extends MastraLLM {
30
- constructor({ name, apiKey, baseURL, }?: {
31
- name?: 'deepseek-r1-distill-llama-70b';
32
- apiKey?: string;
33
- baseURL?: string;
34
- });
35
- }
36
-
37
- export { Groq, type GroqModel, GroqReasoning };
@@ -1,42 +0,0 @@
1
- import { MastraLLM } from '../../../chunk-QX2CHXA7.js';
2
- import '../../../chunk-KNPBNSJ7.js';
3
- import '../../../chunk-G4MCO7XF.js';
4
- import '../../../chunk-ICMEXHKD.js';
5
- import { __name } from '../../../chunk-AJJZUHB4.js';
6
- import { createGroq } from '@ai-sdk/groq';
7
- import { wrapLanguageModel, extractReasoningMiddleware } from 'ai';
8
-
9
- var _Groq = class _Groq extends MastraLLM {
10
- constructor({ name = "gemma2-9b-it", apiKey = process.env.GROQ_API_KEY || "", baseURL = "https://api.groq.com/openai/v1" } = {}) {
11
- const groqModel = createGroq({
12
- baseURL,
13
- apiKey
14
- });
15
- super({
16
- model: groqModel(name)
17
- });
18
- }
19
- };
20
- __name(_Groq, "Groq");
21
- var Groq = _Groq;
22
- var _GroqReasoning = class _GroqReasoning extends MastraLLM {
23
- constructor({ name = "deepseek-r1-distill-llama-70b", apiKey = process.env.GROQ_API_KEY || "", baseURL = "https://api.groq.com/openai/v1" } = {}) {
24
- const groqModel = createGroq({
25
- baseURL,
26
- apiKey
27
- });
28
- const enhancedModel = wrapLanguageModel({
29
- model: groqModel(name),
30
- middleware: extractReasoningMiddleware({
31
- tagName: "think"
32
- })
33
- });
34
- super({
35
- model: enhancedModel
36
- });
37
- }
38
- };
39
- __name(_GroqReasoning, "GroqReasoning");
40
- var GroqReasoning = _GroqReasoning;
41
-
42
- export { Groq, GroqReasoning };
@@ -1,29 +0,0 @@
1
- import { MastraLLM } from '../index.js';
2
- import '../../../index-B48181D5.js';
3
- import 'ai';
4
- import 'json-schema';
5
- import 'zod';
6
- import '../../../base.js';
7
- import '@opentelemetry/api';
8
- import '../../../index-CBZ2mk2H.js';
9
- import 'pino';
10
- import 'stream';
11
- import '../../../telemetry-oCUM52DG.js';
12
- import '@opentelemetry/sdk-node';
13
- import '@opentelemetry/sdk-trace-base';
14
- import '../../../metric-D2V4CR8D.js';
15
- import 'sift';
16
- import '../../../embeddings/model/providers/embedder.js';
17
- import '../../../vector/index.js';
18
- import '../../../types-M16hSruO.js';
19
- import '../../../tts/index.js';
20
-
21
- type LMStudioModel = 'qwen2-7b-instruct-4bit' | 'qwen2-math-1.5b' | 'qwen2-0.5b' | 'aya-23-8b' | 'mistral-7b-v0.3' | 'stablecode' | 'cohere-command-r-v01-4bit' | 'command-r' | 'starcoder2-7b' | 'deepseek-math-7b' | 'qwen2.5-coder-14b' | 'qwen2.5-coder-32b' | 'qwen2.5-coder-3b' | 'llama-3.2-3b-instruct-4bit' | 'llama-3.2-1b' | 'llama-3.2-3b' | 'qwen2.5-coder-7b' | 'qwen2.5-14b' | 'yi-coder-9b' | 'hermes-3-llama-3.1-8b' | 'internlm-2.5-20b' | 'llava-v1.5' | 'llama-3.1-8b-instruct-4bit' | 'meta-llama-3.1-8b' | 'mistral-nemo-2407' | 'mistral-nemo-instruct-2407-4bit' | 'gemma-2-2b' | 'mathstral-7b' | 'gemma-2-9b' | 'deepseek-coder-v2-lite-instruct-4bit' | 'smollm-360m-v0.2' | 'phi-3-mini-4k-instruct-4bit' | 'gemma-2-27b' | 'codestral-22b' | 'phi-3.1-mini-128k' | 'deepseek-coder-v2-lite' | (string & {});
22
- declare class LMStudio extends MastraLLM {
23
- constructor({ name, baseURL }: {
24
- name: LMStudioModel;
25
- baseURL: string;
26
- });
27
- }
28
-
29
- export { LMStudio, type LMStudioModel };
@@ -1,22 +0,0 @@
1
- import { openaiCompat } from '../../../chunk-NYBGZL6Z.js';
2
- import { MastraLLM } from '../../../chunk-QX2CHXA7.js';
3
- import '../../../chunk-KNPBNSJ7.js';
4
- import '../../../chunk-G4MCO7XF.js';
5
- import '../../../chunk-ICMEXHKD.js';
6
- import { __name } from '../../../chunk-AJJZUHB4.js';
7
-
8
- // src/llm/model/providers/lmstudio.ts
9
- var _LMStudio = class _LMStudio extends MastraLLM {
10
- constructor({ name, baseURL }) {
11
- super({
12
- model: openaiCompat({
13
- modelName: name,
14
- baseURL
15
- })
16
- });
17
- }
18
- };
19
- __name(_LMStudio, "LMStudio");
20
- var LMStudio = _LMStudio;
21
-
22
- export { LMStudio };
@@ -1,30 +0,0 @@
1
- import { MastraLLM } from '../index.js';
2
- import '../../../index-B48181D5.js';
3
- import 'ai';
4
- import 'json-schema';
5
- import 'zod';
6
- import '../../../base.js';
7
- import '@opentelemetry/api';
8
- import '../../../index-CBZ2mk2H.js';
9
- import 'pino';
10
- import 'stream';
11
- import '../../../telemetry-oCUM52DG.js';
12
- import '@opentelemetry/sdk-node';
13
- import '@opentelemetry/sdk-trace-base';
14
- import '../../../metric-D2V4CR8D.js';
15
- import 'sift';
16
- import '../../../embeddings/model/providers/embedder.js';
17
- import '../../../vector/index.js';
18
- import '../../../types-M16hSruO.js';
19
- import '../../../tts/index.js';
20
-
21
- type MistralModel = 'pixtral-large-latest' | 'mistral-large-latest' | 'mistral-small-latest' | 'ministral-3b-latest' | 'ministral-8b-latest' | 'pixtral-12b-2409' | (string & {});
22
- declare class Mistral extends MastraLLM {
23
- constructor({ name, apiKey, baseURL, }?: {
24
- name?: MistralModel;
25
- apiKey?: string;
26
- baseURL?: string;
27
- });
28
- }
29
-
30
- export { Mistral, type MistralModel };
@@ -1,22 +0,0 @@
1
- import { MastraLLM } from '../../../chunk-QX2CHXA7.js';
2
- import '../../../chunk-KNPBNSJ7.js';
3
- import '../../../chunk-G4MCO7XF.js';
4
- import '../../../chunk-ICMEXHKD.js';
5
- import { __name } from '../../../chunk-AJJZUHB4.js';
6
- import { createMistral } from '@ai-sdk/mistral';
7
-
8
- var _Mistral = class _Mistral extends MastraLLM {
9
- constructor({ name = "pixtral-large-latest", apiKey = process.env.MISTRAL_API_KEY || "", baseURL = "https://api.mistral.ai/v1" } = {}) {
10
- const mistralModel = createMistral({
11
- baseURL,
12
- apiKey
13
- });
14
- super({
15
- model: mistralModel(name)
16
- });
17
- }
18
- };
19
- __name(_Mistral, "Mistral");
20
- var Mistral = _Mistral;
21
-
22
- export { Mistral };
@@ -1,30 +0,0 @@
1
- import { MastraLLM } from '../index.js';
2
- import '../../../index-B48181D5.js';
3
- import 'ai';
4
- import 'json-schema';
5
- import 'zod';
6
- import '../../../base.js';
7
- import '@opentelemetry/api';
8
- import '../../../index-CBZ2mk2H.js';
9
- import 'pino';
10
- import 'stream';
11
- import '../../../telemetry-oCUM52DG.js';
12
- import '@opentelemetry/sdk-node';
13
- import '@opentelemetry/sdk-trace-base';
14
- import '../../../metric-D2V4CR8D.js';
15
- import 'sift';
16
- import '../../../embeddings/model/providers/embedder.js';
17
- import '../../../vector/index.js';
18
- import '../../../types-M16hSruO.js';
19
- import '../../../tts/index.js';
20
-
21
- declare class MockProvider extends MastraLLM {
22
- constructor({ spyGenerate, spyStream, objectGenerationMode, mockText, }: {
23
- spyGenerate?: (props: any) => void;
24
- spyStream?: (props: any) => void;
25
- objectGenerationMode?: 'json';
26
- mockText?: string | Record<string, any>;
27
- });
28
- }
29
-
30
- export { MockProvider };
@@ -1,83 +0,0 @@
1
- import { MastraLLM } from '../../../chunk-QX2CHXA7.js';
2
- import '../../../chunk-KNPBNSJ7.js';
3
- import '../../../chunk-G4MCO7XF.js';
4
- import '../../../chunk-ICMEXHKD.js';
5
- import { __name } from '../../../chunk-AJJZUHB4.js';
6
- import { simulateReadableStream } from 'ai';
7
- import { MockLanguageModelV1 } from 'ai/test';
8
-
9
- var _MockProvider = class _MockProvider extends MastraLLM {
10
- constructor({ spyGenerate, spyStream, objectGenerationMode, mockText = "Hello, world!" }) {
11
- const mockModel = new MockLanguageModelV1({
12
- defaultObjectGenerationMode: objectGenerationMode,
13
- doGenerate: /* @__PURE__ */ __name(async (props) => {
14
- if (spyGenerate) {
15
- spyGenerate(props);
16
- }
17
- if (objectGenerationMode === "json") {
18
- return {
19
- rawCall: {
20
- rawPrompt: null,
21
- rawSettings: {}
22
- },
23
- finishReason: "stop",
24
- usage: {
25
- promptTokens: 10,
26
- completionTokens: 20
27
- },
28
- text: JSON.stringify(mockText)
29
- };
30
- }
31
- return {
32
- rawCall: {
33
- rawPrompt: null,
34
- rawSettings: {}
35
- },
36
- finishReason: "stop",
37
- usage: {
38
- promptTokens: 10,
39
- completionTokens: 20
40
- },
41
- text: typeof mockText === "string" ? mockText : JSON.stringify(mockText)
42
- };
43
- }, "doGenerate"),
44
- doStream: /* @__PURE__ */ __name(async (props) => {
45
- if (spyStream) {
46
- spyStream(props);
47
- }
48
- const text = typeof mockText === "string" ? mockText : JSON.stringify(mockText);
49
- const chunks = text.split(" ").map((word) => ({
50
- type: "text-delta",
51
- textDelta: word + " "
52
- }));
53
- return {
54
- stream: simulateReadableStream({
55
- chunks: [
56
- ...chunks,
57
- {
58
- type: "finish",
59
- finishReason: "stop",
60
- logprobs: undefined,
61
- usage: {
62
- completionTokens: 10,
63
- promptTokens: 3
64
- }
65
- }
66
- ]
67
- }),
68
- rawCall: {
69
- rawPrompt: null,
70
- rawSettings: {}
71
- }
72
- };
73
- }, "doStream")
74
- });
75
- super({
76
- model: mockModel
77
- });
78
- }
79
- };
80
- __name(_MockProvider, "MockProvider");
81
- var MockProvider = _MockProvider;
82
-
83
- export { MockProvider };
@@ -1,31 +0,0 @@
1
- import { MastraLLM } from '../index.js';
2
- import '../../../index-B48181D5.js';
3
- import 'ai';
4
- import 'json-schema';
5
- import 'zod';
6
- import '../../../base.js';
7
- import '@opentelemetry/api';
8
- import '../../../index-CBZ2mk2H.js';
9
- import 'pino';
10
- import 'stream';
11
- import '../../../telemetry-oCUM52DG.js';
12
- import '@opentelemetry/sdk-node';
13
- import '@opentelemetry/sdk-trace-base';
14
- import '../../../metric-D2V4CR8D.js';
15
- import 'sift';
16
- import '../../../embeddings/model/providers/embedder.js';
17
- import '../../../vector/index.js';
18
- import '../../../types-M16hSruO.js';
19
- import '../../../tts/index.js';
20
-
21
- type OllamaModel = string & {};
22
- declare class OllamaAI extends MastraLLM {
23
- constructor({ name, baseURL, headers, fetch, }: {
24
- name: OllamaModel;
25
- baseURL: string;
26
- headers?: Record<string, string>;
27
- fetch?: typeof globalThis.fetch;
28
- });
29
- }
30
-
31
- export { OllamaAI, type OllamaModel };
@@ -1,23 +0,0 @@
1
- import { MastraLLM } from '../../../chunk-QX2CHXA7.js';
2
- import '../../../chunk-KNPBNSJ7.js';
3
- import '../../../chunk-G4MCO7XF.js';
4
- import '../../../chunk-ICMEXHKD.js';
5
- import { __name } from '../../../chunk-AJJZUHB4.js';
6
- import { createOllama } from 'ollama-ai-provider';
7
-
8
- var _OllamaAI = class _OllamaAI extends MastraLLM {
9
- constructor({ name, baseURL, headers, fetch }) {
10
- const ollama = createOllama({
11
- baseURL,
12
- fetch,
13
- headers
14
- });
15
- super({
16
- model: ollama(name)
17
- });
18
- }
19
- };
20
- __name(_OllamaAI, "OllamaAI");
21
- var OllamaAI = _OllamaAI;
22
-
23
- export { OllamaAI };
@@ -1,39 +0,0 @@
1
- import * as ai from 'ai';
2
- import { OpenAIChatSettings } from '@ai-sdk/openai/internal';
3
- import { MastraLLM } from '../index.js';
4
- import '../../../index-B48181D5.js';
5
- import 'json-schema';
6
- import 'zod';
7
- import '../../../base.js';
8
- import '@opentelemetry/api';
9
- import '../../../index-CBZ2mk2H.js';
10
- import 'pino';
11
- import 'stream';
12
- import '../../../telemetry-oCUM52DG.js';
13
- import '@opentelemetry/sdk-node';
14
- import '@opentelemetry/sdk-trace-base';
15
- import '../../../metric-D2V4CR8D.js';
16
- import 'sift';
17
- import '../../../embeddings/model/providers/embedder.js';
18
- import '../../../vector/index.js';
19
- import '../../../types-M16hSruO.js';
20
- import '../../../tts/index.js';
21
-
22
- declare function openaiCompat({ baseURL, apiKey, modelName, fetch, settings, }: {
23
- baseURL?: string;
24
- apiKey?: string;
25
- modelName: string;
26
- fetch?: typeof globalThis.fetch;
27
- settings?: OpenAIChatSettings;
28
- }): ai.LanguageModelV1;
29
- declare class OpenaiCompat extends MastraLLM {
30
- constructor({ name, apiKey, baseURL, fetch, settings, }: {
31
- name: string;
32
- apiKey?: string;
33
- baseURL?: string;
34
- fetch?: typeof globalThis.fetch;
35
- settings?: OpenAIChatSettings;
36
- });
37
- }
38
-
39
- export { OpenaiCompat, openaiCompat };
@@ -1,6 +0,0 @@
1
- export { OpenaiCompat, openaiCompat } from '../../../chunk-NYBGZL6Z.js';
2
- import '../../../chunk-QX2CHXA7.js';
3
- import '../../../chunk-KNPBNSJ7.js';
4
- import '../../../chunk-G4MCO7XF.js';
5
- import '../../../chunk-ICMEXHKD.js';
6
- import '../../../chunk-AJJZUHB4.js';