langchain 0.0.137 → 0.0.138

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -29,6 +29,42 @@ export class Ollama extends LLM {
29
29
  writable: true,
30
30
  value: "http://localhost:11434"
31
31
  });
32
+ Object.defineProperty(this, "embeddingOnly", {
33
+ enumerable: true,
34
+ configurable: true,
35
+ writable: true,
36
+ value: void 0
37
+ });
38
+ Object.defineProperty(this, "f16KV", {
39
+ enumerable: true,
40
+ configurable: true,
41
+ writable: true,
42
+ value: void 0
43
+ });
44
+ Object.defineProperty(this, "frequencyPenalty", {
45
+ enumerable: true,
46
+ configurable: true,
47
+ writable: true,
48
+ value: void 0
49
+ });
50
+ Object.defineProperty(this, "logitsAll", {
51
+ enumerable: true,
52
+ configurable: true,
53
+ writable: true,
54
+ value: void 0
55
+ });
56
+ Object.defineProperty(this, "lowVram", {
57
+ enumerable: true,
58
+ configurable: true,
59
+ writable: true,
60
+ value: void 0
61
+ });
62
+ Object.defineProperty(this, "mainGpu", {
63
+ enumerable: true,
64
+ configurable: true,
65
+ writable: true,
66
+ value: void 0
67
+ });
32
68
  Object.defineProperty(this, "mirostat", {
33
69
  enumerable: true,
34
70
  configurable: true,
@@ -47,6 +83,12 @@ export class Ollama extends LLM {
47
83
  writable: true,
48
84
  value: void 0
49
85
  });
86
+ Object.defineProperty(this, "numBatch", {
87
+ enumerable: true,
88
+ configurable: true,
89
+ writable: true,
90
+ value: void 0
91
+ });
50
92
  Object.defineProperty(this, "numCtx", {
51
93
  enumerable: true,
52
94
  configurable: true,
@@ -59,12 +101,36 @@ export class Ollama extends LLM {
59
101
  writable: true,
60
102
  value: void 0
61
103
  });
104
+ Object.defineProperty(this, "numGqa", {
105
+ enumerable: true,
106
+ configurable: true,
107
+ writable: true,
108
+ value: void 0
109
+ });
110
+ Object.defineProperty(this, "numKeep", {
111
+ enumerable: true,
112
+ configurable: true,
113
+ writable: true,
114
+ value: void 0
115
+ });
62
116
  Object.defineProperty(this, "numThread", {
63
117
  enumerable: true,
64
118
  configurable: true,
65
119
  writable: true,
66
120
  value: void 0
67
121
  });
122
+ Object.defineProperty(this, "penalizeNewline", {
123
+ enumerable: true,
124
+ configurable: true,
125
+ writable: true,
126
+ value: void 0
127
+ });
128
+ Object.defineProperty(this, "presencePenalty", {
129
+ enumerable: true,
130
+ configurable: true,
131
+ writable: true,
132
+ value: void 0
133
+ });
68
134
  Object.defineProperty(this, "repeatLastN", {
69
135
  enumerable: true,
70
136
  configurable: true,
@@ -77,6 +143,18 @@ export class Ollama extends LLM {
77
143
  writable: true,
78
144
  value: void 0
79
145
  });
146
+ Object.defineProperty(this, "ropeFrequencyBase", {
147
+ enumerable: true,
148
+ configurable: true,
149
+ writable: true,
150
+ value: void 0
151
+ });
152
+ Object.defineProperty(this, "ropeFrequencyScale", {
153
+ enumerable: true,
154
+ configurable: true,
155
+ writable: true,
156
+ value: void 0
157
+ });
80
158
  Object.defineProperty(this, "temperature", {
81
159
  enumerable: true,
82
160
  configurable: true,
@@ -107,23 +185,64 @@ export class Ollama extends LLM {
107
185
  writable: true,
108
186
  value: void 0
109
187
  });
188
+ Object.defineProperty(this, "typicalP", {
189
+ enumerable: true,
190
+ configurable: true,
191
+ writable: true,
192
+ value: void 0
193
+ });
194
+ Object.defineProperty(this, "useMLock", {
195
+ enumerable: true,
196
+ configurable: true,
197
+ writable: true,
198
+ value: void 0
199
+ });
200
+ Object.defineProperty(this, "useMMap", {
201
+ enumerable: true,
202
+ configurable: true,
203
+ writable: true,
204
+ value: void 0
205
+ });
206
+ Object.defineProperty(this, "vocabOnly", {
207
+ enumerable: true,
208
+ configurable: true,
209
+ writable: true,
210
+ value: void 0
211
+ });
110
212
  this.model = fields.model ?? this.model;
111
213
  this.baseUrl = fields.baseUrl?.endsWith("/")
112
214
  ? fields.baseUrl.slice(0, -1)
113
215
  : fields.baseUrl ?? this.baseUrl;
216
+ this.embeddingOnly = fields.embeddingOnly;
217
+ this.f16KV = fields.f16KV;
218
+ this.frequencyPenalty = fields.frequencyPenalty;
219
+ this.logitsAll = fields.logitsAll;
220
+ this.lowVram = fields.lowVram;
221
+ this.mainGpu = fields.mainGpu;
114
222
  this.mirostat = fields.mirostat;
115
223
  this.mirostatEta = fields.mirostatEta;
116
224
  this.mirostatTau = fields.mirostatTau;
225
+ this.numBatch = fields.numBatch;
117
226
  this.numCtx = fields.numCtx;
118
227
  this.numGpu = fields.numGpu;
228
+ this.numGqa = fields.numGqa;
229
+ this.numKeep = fields.numKeep;
119
230
  this.numThread = fields.numThread;
231
+ this.penalizeNewline = fields.penalizeNewline;
232
+ this.presencePenalty = fields.presencePenalty;
120
233
  this.repeatLastN = fields.repeatLastN;
121
234
  this.repeatPenalty = fields.repeatPenalty;
235
+ this.ropeFrequencyBase = fields.ropeFrequencyBase;
236
+ this.ropeFrequencyScale = fields.ropeFrequencyScale;
122
237
  this.temperature = fields.temperature;
123
238
  this.stop = fields.stop;
124
239
  this.tfsZ = fields.tfsZ;
125
240
  this.topK = fields.topK;
126
241
  this.topP = fields.topP;
242
+ this.typicalP = fields.typicalP;
243
+ this.useMLock = fields.useMLock;
244
+ this.useMMap = fields.useMMap;
245
+ this.vocabOnly = fields.vocabOnly;
127
246
  }
128
247
  _llmType() {
129
248
  return "ollama";
@@ -132,19 +251,36 @@ export class Ollama extends LLM {
132
251
  return {
133
252
  model: this.model,
134
253
  options: {
254
+ embedding_only: this.embeddingOnly,
255
+ f16_kv: this.f16KV,
256
+ frequency_penalty: this.frequencyPenalty,
257
+ logits_all: this.logitsAll,
258
+ low_vram: this.lowVram,
259
+ main_gpu: this.mainGpu,
135
260
  mirostat: this.mirostat,
136
261
  mirostat_eta: this.mirostatEta,
137
262
  mirostat_tau: this.mirostatTau,
263
+ num_batch: this.numBatch,
138
264
  num_ctx: this.numCtx,
139
265
  num_gpu: this.numGpu,
266
+ num_gqa: this.numGqa,
267
+ num_keep: this.numKeep,
140
268
  num_thread: this.numThread,
269
+ penalize_newline: this.penalizeNewline,
270
+ presence_penalty: this.presencePenalty,
141
271
  repeat_last_n: this.repeatLastN,
142
272
  repeat_penalty: this.repeatPenalty,
273
+ rope_frequency_base: this.ropeFrequencyBase,
274
+ rope_frequency_scale: this.ropeFrequencyScale,
143
275
  temperature: this.temperature,
144
276
  stop: options?.stop ?? this.stop,
145
277
  tfs_z: this.tfsZ,
146
278
  top_k: this.topK,
147
279
  top_p: this.topP,
280
+ typical_p: this.typicalP,
281
+ use_mlock: this.useMLock,
282
+ use_mmap: this.useMMap,
283
+ vocab_only: this.vocabOnly,
148
284
  },
149
285
  };
150
286
  }
@@ -24,8 +24,8 @@ var __importStar = (this && this.__importStar) || function (mod) {
24
24
  return result;
25
25
  };
26
26
  Object.defineProperty(exports, "__esModule", { value: true });
27
- exports.retrievers__document_compressors__chain_extract = exports.retrievers__time_weighted = exports.retrievers__parent_document = exports.retrievers__multi_vector = exports.retrievers__document_compressors = exports.retrievers__contextual_compression = exports.retrievers__databerry = exports.retrievers__remote = exports.output_parsers = exports.callbacks = exports.schema__storage = exports.schema__runnable = exports.schema__retriever = exports.schema__query_constructor = exports.schema__output_parser = exports.schema = exports.chat_models__ollama = exports.chat_models__baiduwenxin = exports.chat_models__anthropic = exports.chat_models__openai = exports.chat_models__base = exports.document_transformers__openai_functions = exports.document_loaders__web__sort_xyz_blockchain = exports.document_loaders__web__serpapi = exports.document_loaders__base = exports.document = exports.memory = exports.text_splitter = exports.vectorstores__xata = exports.vectorstores__vectara = exports.vectorstores__prisma = exports.vectorstores__memory = exports.vectorstores__base = exports.prompts = exports.llms__ollama = exports.llms__aleph_alpha = exports.llms__ai21 = exports.llms__openai = exports.llms__base = exports.embeddings__openai = exports.embeddings__fake = exports.embeddings__cache_backed = exports.embeddings__base = exports.chains__openai_functions = exports.chains = exports.tools = exports.base_language = exports.agents__toolkits = exports.agents = exports.load__serializable = void 0;
28
- exports.evaluation = exports.experimental__plan_and_execute = exports.experimental__generative_agents = exports.experimental__babyagi = exports.experimental__autogpt = exports.util__math = exports.storage__in_memory = exports.stores__message__in_memory = exports.stores__file__in_memory = exports.stores__doc__in_memory = exports.cache = exports.retrievers__vespa = exports.retrievers__score_threshold = exports.retrievers__hyde = void 0;
27
+ exports.retrievers__parent_document = exports.retrievers__multi_vector = exports.retrievers__document_compressors = exports.retrievers__contextual_compression = exports.retrievers__databerry = exports.retrievers__remote = exports.output_parsers = exports.callbacks = exports.schema__storage = exports.schema__runnable = exports.schema__retriever = exports.schema__query_constructor = exports.schema__output_parser = exports.schema = exports.chat_models__minimax = exports.chat_models__ollama = exports.chat_models__baiduwenxin = exports.chat_models__anthropic = exports.chat_models__openai = exports.chat_models__base = exports.document_transformers__openai_functions = exports.document_loaders__web__sort_xyz_blockchain = exports.document_loaders__web__serpapi = exports.document_loaders__base = exports.document = exports.memory = exports.text_splitter = exports.vectorstores__xata = exports.vectorstores__vectara = exports.vectorstores__prisma = exports.vectorstores__memory = exports.vectorstores__base = exports.prompts = exports.llms__ollama = exports.llms__aleph_alpha = exports.llms__ai21 = exports.llms__openai = exports.llms__base = exports.embeddings__minimax = exports.embeddings__openai = exports.embeddings__fake = exports.embeddings__cache_backed = exports.embeddings__base = exports.chains__openai_functions = exports.chains = exports.tools = exports.base_language = exports.agents__toolkits = exports.agents = exports.load__serializable = void 0;
28
+ exports.evaluation = exports.experimental__plan_and_execute = exports.experimental__generative_agents = exports.experimental__babyagi = exports.experimental__autogpt = exports.util__math = exports.storage__in_memory = exports.stores__message__in_memory = exports.stores__file__in_memory = exports.stores__doc__in_memory = exports.cache = exports.retrievers__vespa = exports.retrievers__score_threshold = exports.retrievers__hyde = exports.retrievers__document_compressors__chain_extract = exports.retrievers__time_weighted = void 0;
29
29
  exports.load__serializable = __importStar(require("../load/serializable.cjs"));
30
30
  exports.agents = __importStar(require("../agents/index.cjs"));
31
31
  exports.agents__toolkits = __importStar(require("../agents/toolkits/index.cjs"));
@@ -37,6 +37,7 @@ exports.embeddings__base = __importStar(require("../embeddings/base.cjs"));
37
37
  exports.embeddings__cache_backed = __importStar(require("../embeddings/cache_backed.cjs"));
38
38
  exports.embeddings__fake = __importStar(require("../embeddings/fake.cjs"));
39
39
  exports.embeddings__openai = __importStar(require("../embeddings/openai.cjs"));
40
+ exports.embeddings__minimax = __importStar(require("../embeddings/minimax.cjs"));
40
41
  exports.llms__base = __importStar(require("../llms/base.cjs"));
41
42
  exports.llms__openai = __importStar(require("../llms/openai.cjs"));
42
43
  exports.llms__ai21 = __importStar(require("../llms/ai21.cjs"));
@@ -60,6 +61,7 @@ exports.chat_models__openai = __importStar(require("../chat_models/openai.cjs"))
60
61
  exports.chat_models__anthropic = __importStar(require("../chat_models/anthropic.cjs"));
61
62
  exports.chat_models__baiduwenxin = __importStar(require("../chat_models/baiduwenxin.cjs"));
62
63
  exports.chat_models__ollama = __importStar(require("../chat_models/ollama.cjs"));
64
+ exports.chat_models__minimax = __importStar(require("../chat_models/minimax.cjs"));
63
65
  exports.schema = __importStar(require("../schema/index.cjs"));
64
66
  exports.schema__output_parser = __importStar(require("../schema/output_parser.cjs"));
65
67
  exports.schema__query_constructor = __importStar(require("../schema/query_constructor.cjs"));
@@ -9,6 +9,7 @@ export * as embeddings__base from "../embeddings/base.js";
9
9
  export * as embeddings__cache_backed from "../embeddings/cache_backed.js";
10
10
  export * as embeddings__fake from "../embeddings/fake.js";
11
11
  export * as embeddings__openai from "../embeddings/openai.js";
12
+ export * as embeddings__minimax from "../embeddings/minimax.js";
12
13
  export * as llms__base from "../llms/base.js";
13
14
  export * as llms__openai from "../llms/openai.js";
14
15
  export * as llms__ai21 from "../llms/ai21.js";
@@ -32,6 +33,7 @@ export * as chat_models__openai from "../chat_models/openai.js";
32
33
  export * as chat_models__anthropic from "../chat_models/anthropic.js";
33
34
  export * as chat_models__baiduwenxin from "../chat_models/baiduwenxin.js";
34
35
  export * as chat_models__ollama from "../chat_models/ollama.js";
36
+ export * as chat_models__minimax from "../chat_models/minimax.js";
35
37
  export * as schema from "../schema/index.js";
36
38
  export * as schema__output_parser from "../schema/output_parser.js";
37
39
  export * as schema__query_constructor from "../schema/query_constructor.js";
@@ -10,6 +10,7 @@ export * as embeddings__base from "../embeddings/base.js";
10
10
  export * as embeddings__cache_backed from "../embeddings/cache_backed.js";
11
11
  export * as embeddings__fake from "../embeddings/fake.js";
12
12
  export * as embeddings__openai from "../embeddings/openai.js";
13
+ export * as embeddings__minimax from "../embeddings/minimax.js";
13
14
  export * as llms__base from "../llms/base.js";
14
15
  export * as llms__openai from "../llms/openai.js";
15
16
  export * as llms__ai21 from "../llms/ai21.js";
@@ -33,6 +34,7 @@ export * as chat_models__openai from "../chat_models/openai.js";
33
34
  export * as chat_models__anthropic from "../chat_models/anthropic.js";
34
35
  export * as chat_models__baiduwenxin from "../chat_models/baiduwenxin.js";
35
36
  export * as chat_models__ollama from "../chat_models/ollama.js";
37
+ export * as chat_models__minimax from "../chat_models/minimax.js";
36
38
  export * as schema from "../schema/index.js";
37
39
  export * as schema__output_parser from "../schema/output_parser.js";
38
40
  export * as schema__query_constructor from "../schema/query_constructor.js";
@@ -1,38 +1,72 @@
1
1
  import { BaseLanguageModelCallOptions } from "../base_language/index.js";
2
2
  export interface OllamaInput {
3
+ embeddingOnly?: boolean;
4
+ f16KV?: boolean;
5
+ frequencyPenalty?: number;
6
+ logitsAll?: boolean;
7
+ lowVram?: boolean;
8
+ mainGpu?: number;
3
9
  model?: string;
4
10
  baseUrl?: string;
5
11
  mirostat?: number;
6
12
  mirostatEta?: number;
7
13
  mirostatTau?: number;
14
+ numBatch?: number;
8
15
  numCtx?: number;
9
16
  numGpu?: number;
17
+ numGqa?: number;
18
+ numKeep?: number;
10
19
  numThread?: number;
20
+ penalizeNewline?: boolean;
21
+ presencePenalty?: number;
11
22
  repeatLastN?: number;
12
23
  repeatPenalty?: number;
24
+ ropeFrequencyBase?: number;
25
+ ropeFrequencyScale?: number;
13
26
  temperature?: number;
14
27
  stop?: string[];
15
28
  tfsZ?: number;
16
29
  topK?: number;
17
30
  topP?: number;
31
+ typicalP?: number;
32
+ useMLock?: boolean;
33
+ useMMap?: boolean;
34
+ vocabOnly?: boolean;
18
35
  }
19
36
  export interface OllamaRequestParams {
20
37
  model: string;
21
38
  prompt: string;
22
39
  options: {
40
+ embedding_only?: boolean;
41
+ f16_kv?: boolean;
42
+ frequency_penalty?: number;
43
+ logits_all?: boolean;
44
+ low_vram?: boolean;
45
+ main_gpu?: number;
23
46
  mirostat?: number;
24
47
  mirostat_eta?: number;
25
48
  mirostat_tau?: number;
49
+ num_batch?: number;
26
50
  num_ctx?: number;
27
51
  num_gpu?: number;
52
+ num_gqa?: number;
53
+ num_keep?: number;
28
54
  num_thread?: number;
55
+ penalize_newline?: boolean;
56
+ presence_penalty?: number;
29
57
  repeat_last_n?: number;
30
58
  repeat_penalty?: number;
59
+ rope_frequency_base?: number;
60
+ rope_frequency_scale?: number;
31
61
  temperature?: number;
32
62
  stop?: string[];
33
63
  tfs_z?: number;
34
64
  top_k?: number;
35
65
  top_p?: number;
66
+ typical_p?: number;
67
+ use_mlock?: boolean;
68
+ use_mmap?: boolean;
69
+ vocab_only?: boolean;
36
70
  };
37
71
  }
38
72
  export interface OllamaCallOptions extends BaseLanguageModelCallOptions {
@@ -235,17 +235,32 @@ class RedisVectorStore extends base_js_1.VectorStore {
235
235
  }
236
236
  /**
237
237
  * Method for dropping an index from the RedisVectorStore.
238
+ * @param deleteDocuments Optional boolean indicating whether to drop the associated documents.
238
239
  * @returns A promise that resolves to a boolean indicating whether the index was dropped.
239
240
  */
240
- async dropIndex() {
241
+ async dropIndex(deleteDocuments) {
241
242
  try {
242
- await this.redisClient.ft.dropIndex(this.indexName);
243
+ const options = deleteDocuments ? { DD: deleteDocuments } : undefined;
244
+ await this.redisClient.ft.dropIndex(this.indexName, options);
243
245
  return true;
244
246
  }
245
247
  catch (err) {
246
248
  return false;
247
249
  }
248
250
  }
251
+ /**
252
+ * Deletes vectors from the vector store.
253
+ * @param params The parameters for deleting vectors.
254
+ * @returns A promise that resolves when the vectors have been deleted.
255
+ */
256
+ async delete(params) {
257
+ if (params.deleteAll) {
258
+ await this.dropIndex(true);
259
+ }
260
+ else {
261
+ throw new Error(`Invalid parameters passed to "delete".`);
262
+ }
263
+ }
249
264
  buildQuery(query, k, filter) {
250
265
  const vectorScoreField = "vector_score";
251
266
  let hybridFields = "*";
@@ -136,9 +136,18 @@ export declare class RedisVectorStore extends VectorStore {
136
136
  createIndex(dimensions?: number): Promise<void>;
137
137
  /**
138
138
  * Method for dropping an index from the RedisVectorStore.
139
+ * @param deleteDocuments Optional boolean indicating whether to drop the associated documents.
139
140
  * @returns A promise that resolves to a boolean indicating whether the index was dropped.
140
141
  */
141
- dropIndex(): Promise<boolean>;
142
+ dropIndex(deleteDocuments?: boolean): Promise<boolean>;
143
+ /**
144
+ * Deletes vectors from the vector store.
145
+ * @param params The parameters for deleting vectors.
146
+ * @returns A promise that resolves when the vectors have been deleted.
147
+ */
148
+ delete(params: {
149
+ deleteAll: boolean;
150
+ }): Promise<void>;
142
151
  private buildQuery;
143
152
  private prepareFilter;
144
153
  /**
@@ -232,17 +232,32 @@ export class RedisVectorStore extends VectorStore {
232
232
  }
233
233
  /**
234
234
  * Method for dropping an index from the RedisVectorStore.
235
+ * @param deleteDocuments Optional boolean indicating whether to drop the associated documents.
235
236
  * @returns A promise that resolves to a boolean indicating whether the index was dropped.
236
237
  */
237
- async dropIndex() {
238
+ async dropIndex(deleteDocuments) {
238
239
  try {
239
- await this.redisClient.ft.dropIndex(this.indexName);
240
+ const options = deleteDocuments ? { DD: deleteDocuments } : undefined;
241
+ await this.redisClient.ft.dropIndex(this.indexName, options);
240
242
  return true;
241
243
  }
242
244
  catch (err) {
243
245
  return false;
244
246
  }
245
247
  }
248
+ /**
249
+ * Deletes vectors from the vector store.
250
+ * @param params The parameters for deleting vectors.
251
+ * @returns A promise that resolves when the vectors have been deleted.
252
+ */
253
+ async delete(params) {
254
+ if (params.deleteAll) {
255
+ await this.dropIndex(true);
256
+ }
257
+ else {
258
+ throw new Error(`Invalid parameters passed to "delete".`);
259
+ }
260
+ }
246
261
  buildQuery(query, k, filter) {
247
262
  const vectorScoreField = "vector_score";
248
263
  let hybridFields = "*";
@@ -76,7 +76,8 @@ class ZepVectorStore extends base_js_1.VectorStore {
76
76
  catch (err) {
77
77
  // eslint-disable-next-line no-instanceof/no-instanceof
78
78
  if (err instanceof Error) {
79
- if (err.name === "NotFoundError") {
79
+ // eslint-disable-next-line no-instanceof/no-instanceof
80
+ if (err instanceof zep_js_1.NotFoundError || err.name === "NotFoundError") {
80
81
  await this.createCollection(args);
81
82
  }
82
83
  else {
@@ -1,4 +1,4 @@
1
- import { ZepClient } from "@getzep/zep-js";
1
+ import { NotFoundError, ZepClient, } from "@getzep/zep-js";
2
2
  import { VectorStore } from "./base.js";
3
3
  import { Document } from "../document.js";
4
4
  import { FakeEmbeddings } from "../embeddings/fake.js";
@@ -73,7 +73,8 @@ export class ZepVectorStore extends VectorStore {
73
73
  catch (err) {
74
74
  // eslint-disable-next-line no-instanceof/no-instanceof
75
75
  if (err instanceof Error) {
76
- if (err.name === "NotFoundError") {
76
+ // eslint-disable-next-line no-instanceof/no-instanceof
77
+ if (err instanceof NotFoundError || err.name === "NotFoundError") {
77
78
  await this.createCollection(args);
78
79
  }
79
80
  else {
@@ -0,0 +1 @@
1
+ module.exports = require('../dist/embeddings/minimax.cjs');
@@ -0,0 +1 @@
1
+ export * from '../dist/embeddings/minimax.js'
@@ -0,0 +1 @@
1
+ export * from '../dist/embeddings/minimax.js'
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "langchain",
3
- "version": "0.0.137",
3
+ "version": "0.0.138",
4
4
  "description": "Typescript bindings for langchain",
5
5
  "type": "module",
6
6
  "engines": {
@@ -100,6 +100,9 @@
100
100
  "embeddings/googlepalm.cjs",
101
101
  "embeddings/googlepalm.js",
102
102
  "embeddings/googlepalm.d.ts",
103
+ "embeddings/minimax.cjs",
104
+ "embeddings/minimax.js",
105
+ "embeddings/minimax.d.ts",
103
106
  "llms.cjs",
104
107
  "llms.js",
105
108
  "llms.d.ts",
@@ -394,6 +397,9 @@
394
397
  "chat_models/ollama.cjs",
395
398
  "chat_models/ollama.js",
396
399
  "chat_models/ollama.d.ts",
400
+ "chat_models/minimax.cjs",
401
+ "chat_models/minimax.js",
402
+ "chat_models/minimax.d.ts",
397
403
  "schema.cjs",
398
404
  "schema.js",
399
405
  "schema.d.ts",
@@ -622,7 +628,7 @@
622
628
  "@elastic/elasticsearch": "^8.4.0",
623
629
  "@faker-js/faker": "^7.6.0",
624
630
  "@getmetal/metal-sdk": "^4.0.0",
625
- "@getzep/zep-js": "^0.6.3",
631
+ "@getzep/zep-js": "^0.7.0",
626
632
  "@gomomento/sdk": "^1.23.0",
627
633
  "@google-ai/generativelanguage": "^0.2.1",
628
634
  "@google-cloud/storage": "^6.10.1",
@@ -736,7 +742,7 @@
736
742
  "@clickhouse/client": "^0.0.14",
737
743
  "@elastic/elasticsearch": "^8.4.0",
738
744
  "@getmetal/metal-sdk": "*",
739
- "@getzep/zep-js": "^0.6.3",
745
+ "@getzep/zep-js": "^0.7.0",
740
746
  "@gomomento/sdk": "^1.23.0",
741
747
  "@google-ai/generativelanguage": "^0.2.1",
742
748
  "@google-cloud/storage": "^6.10.1",
@@ -1231,6 +1237,11 @@
1231
1237
  "import": "./embeddings/googlepalm.js",
1232
1238
  "require": "./embeddings/googlepalm.cjs"
1233
1239
  },
1240
+ "./embeddings/minimax": {
1241
+ "types": "./embeddings/minimax.d.ts",
1242
+ "import": "./embeddings/minimax.js",
1243
+ "require": "./embeddings/minimax.cjs"
1244
+ },
1234
1245
  "./llms": {
1235
1246
  "node": {
1236
1247
  "types": "./llms.d.ts",
@@ -1729,6 +1740,11 @@
1729
1740
  "import": "./chat_models/ollama.js",
1730
1741
  "require": "./chat_models/ollama.cjs"
1731
1742
  },
1743
+ "./chat_models/minimax": {
1744
+ "types": "./chat_models/minimax.d.ts",
1745
+ "import": "./chat_models/minimax.js",
1746
+ "require": "./chat_models/minimax.cjs"
1747
+ },
1732
1748
  "./schema": {
1733
1749
  "types": "./schema.d.ts",
1734
1750
  "import": "./schema.js",