@copilotkit/runtime 1.6.0-next.1 → 1.6.0-next.11

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (61) hide show
  1. package/CHANGELOG.md +75 -0
  2. package/__snapshots__/schema/schema.graphql +1 -0
  3. package/dist/{chunk-CCQ73DAH.mjs → chunk-6L2NPMPT.mjs} +2 -2
  4. package/dist/{chunk-CHDIEE43.mjs → chunk-DVFBTXBT.mjs} +87 -44
  5. package/dist/chunk-DVFBTXBT.mjs.map +1 -0
  6. package/dist/{chunk-OS5YD32G.mjs → chunk-FZJAYGIR.mjs} +45 -12
  7. package/dist/chunk-FZJAYGIR.mjs.map +1 -0
  8. package/dist/{chunk-C7GTLEVO.mjs → chunk-LGQL6LK4.mjs} +2 -2
  9. package/dist/{chunk-A25FIW7J.mjs → chunk-O3Q4BRZA.mjs} +2 -2
  10. package/dist/{copilot-runtime-67033bfa.d.ts → copilot-runtime-15bfc4f4.d.ts} +2 -2
  11. package/dist/graphql/types/converted/index.d.ts +1 -1
  12. package/dist/{groq-adapter-9d15c927.d.ts → groq-adapter-fb9aa3ab.d.ts} +1 -1
  13. package/dist/{index-f6d1f30b.d.ts → index-5bec5424.d.ts} +2 -1
  14. package/dist/index.d.ts +4 -4
  15. package/dist/index.js +199 -123
  16. package/dist/index.js.map +1 -1
  17. package/dist/index.mjs +5 -5
  18. package/dist/{langserve-7cc5be48.d.ts → langserve-6f7af8d3.d.ts} +1 -1
  19. package/dist/lib/index.d.ts +4 -4
  20. package/dist/lib/index.js +192 -116
  21. package/dist/lib/index.js.map +1 -1
  22. package/dist/lib/index.mjs +5 -5
  23. package/dist/lib/integrations/index.d.ts +4 -4
  24. package/dist/lib/integrations/index.js +5 -3
  25. package/dist/lib/integrations/index.js.map +1 -1
  26. package/dist/lib/integrations/index.mjs +5 -5
  27. package/dist/lib/integrations/nest/index.d.ts +3 -3
  28. package/dist/lib/integrations/nest/index.js +5 -3
  29. package/dist/lib/integrations/nest/index.js.map +1 -1
  30. package/dist/lib/integrations/nest/index.mjs +3 -3
  31. package/dist/lib/integrations/node-express/index.d.ts +3 -3
  32. package/dist/lib/integrations/node-express/index.js +5 -3
  33. package/dist/lib/integrations/node-express/index.js.map +1 -1
  34. package/dist/lib/integrations/node-express/index.mjs +3 -3
  35. package/dist/lib/integrations/node-http/index.d.ts +3 -3
  36. package/dist/lib/integrations/node-http/index.js +5 -3
  37. package/dist/lib/integrations/node-http/index.js.map +1 -1
  38. package/dist/lib/integrations/node-http/index.mjs +2 -2
  39. package/dist/service-adapters/index.d.ts +4 -4
  40. package/dist/service-adapters/index.js +70 -37
  41. package/dist/service-adapters/index.js.map +1 -1
  42. package/dist/service-adapters/index.mjs +1 -1
  43. package/package.json +5 -4
  44. package/src/agents/langgraph/event-source.ts +21 -5
  45. package/src/graphql/types/enums.ts +1 -0
  46. package/src/lib/runtime/__tests__/remote-action-constructors.test.ts +236 -0
  47. package/src/lib/runtime/copilot-runtime.ts +9 -3
  48. package/src/lib/runtime/remote-action-constructors.ts +9 -7
  49. package/src/lib/runtime/remote-lg-action.ts +35 -7
  50. package/src/service-adapters/conversion.ts +39 -46
  51. package/src/service-adapters/groq/groq-adapter.ts +6 -3
  52. package/src/service-adapters/openai/openai-adapter.ts +1 -1
  53. package/src/service-adapters/openai/openai-assistant-adapter.ts +1 -1
  54. package/src/service-adapters/openai/utils.ts +39 -13
  55. package/src/service-adapters/unify/unify-adapter.ts +1 -1
  56. package/tsconfig.json +3 -2
  57. package/dist/chunk-CHDIEE43.mjs.map +0 -1
  58. package/dist/chunk-OS5YD32G.mjs.map +0 -1
  59. /package/dist/{chunk-CCQ73DAH.mjs.map → chunk-6L2NPMPT.mjs.map} +0 -0
  60. /package/dist/{chunk-C7GTLEVO.mjs.map → chunk-LGQL6LK4.mjs.map} +0 -0
  61. /package/dist/{chunk-A25FIW7J.mjs.map → chunk-O3Q4BRZA.mjs.map} +0 -0
@@ -1,7 +1,7 @@
1
1
  import {
2
2
  copilotRuntimeNodeHttpEndpoint
3
- } from "../../../chunk-CHDIEE43.mjs";
4
- import "../../../chunk-OS5YD32G.mjs";
3
+ } from "../../../chunk-DVFBTXBT.mjs";
4
+ import "../../../chunk-FZJAYGIR.mjs";
5
5
  import "../../../chunk-5BIEM2UU.mjs";
6
6
  import "../../../chunk-RTFJTJMA.mjs";
7
7
  import "../../../chunk-2OZAGFV3.mjs";
@@ -1,8 +1,8 @@
1
- import { b as CopilotServiceAdapter, C as CopilotRuntimeChatCompletionRequest, a as CopilotRuntimeChatCompletionResponse } from '../langserve-7cc5be48.js';
2
- export { c as RemoteChain, R as RemoteChainParameters } from '../langserve-7cc5be48.js';
3
- export { G as GoogleGenerativeAIAdapter, f as GroqAdapter, e as GroqAdapterParams, L as LangChainAdapter, a as OpenAIAdapter, O as OpenAIAdapterParams, c as OpenAIAssistantAdapter, b as OpenAIAssistantAdapterParams, d as UnifyAdapter, U as UnifyAdapterParams } from '../groq-adapter-9d15c927.js';
1
+ import { b as CopilotServiceAdapter, C as CopilotRuntimeChatCompletionRequest, a as CopilotRuntimeChatCompletionResponse } from '../langserve-6f7af8d3.js';
2
+ export { c as RemoteChain, R as RemoteChainParameters } from '../langserve-6f7af8d3.js';
3
+ export { G as GoogleGenerativeAIAdapter, f as GroqAdapter, e as GroqAdapterParams, L as LangChainAdapter, a as OpenAIAdapter, O as OpenAIAdapterParams, c as OpenAIAssistantAdapter, b as OpenAIAssistantAdapterParams, d as UnifyAdapter, U as UnifyAdapterParams } from '../groq-adapter-fb9aa3ab.js';
4
4
  import Anthropic from '@anthropic-ai/sdk';
5
- import '../index-f6d1f30b.js';
5
+ import '../index-5bec5424.js';
6
6
  import '../graphql/types/base/index.js';
7
7
  import 'rxjs';
8
8
  import '@copilotkit/shared';
@@ -126,6 +126,7 @@ __name(RemoteChain, "RemoteChain");
126
126
  var import_openai = __toESM(require("openai"));
127
127
 
128
128
  // src/service-adapters/openai/utils.ts
129
+ var import_shared = require("@copilotkit/shared");
129
130
  function limitMessagesToTokenCount(messages, tools, model, maxTokens) {
130
131
  maxTokens || (maxTokens = maxTokensForOpenAIModel(model));
131
132
  const result = [];
@@ -135,7 +136,10 @@ function limitMessagesToTokenCount(messages, tools, model, maxTokens) {
135
136
  }
136
137
  maxTokens -= toolsNumTokens;
137
138
  for (const message of messages) {
138
- if (message.role === "system") {
139
+ if ([
140
+ "system",
141
+ "developer"
142
+ ].includes(message.role)) {
139
143
  const numTokens = countMessageTokens(model, message);
140
144
  maxTokens -= numTokens;
141
145
  if (maxTokens < 0) {
@@ -148,7 +152,10 @@ function limitMessagesToTokenCount(messages, tools, model, maxTokens) {
148
152
  ...messages
149
153
  ].reverse();
150
154
  for (const message of reversedMessages) {
151
- if (message.role === "system") {
155
+ if ([
156
+ "system",
157
+ "developer"
158
+ ].includes(message.role)) {
152
159
  result.unshift(message);
153
160
  continue;
154
161
  } else if (cutoff) {
@@ -171,9 +178,23 @@ function maxTokensForOpenAIModel(model) {
171
178
  __name(maxTokensForOpenAIModel, "maxTokensForOpenAIModel");
172
179
  var DEFAULT_MAX_TOKENS = 128e3;
173
180
  var maxTokensByModel = {
181
+ // o1
182
+ o1: 2e5,
183
+ "o1-2024-12-17": 2e5,
184
+ "o1-mini": 128e3,
185
+ "o1-mini-2024-09-12": 128e3,
186
+ "o1-preview": 128e3,
187
+ "o1-preview-2024-09-12": 128e3,
188
+ // o3-mini
189
+ "o3-mini": 2e5,
190
+ "o3-mini-2025-01-31": 2e5,
174
191
  // GPT-4
175
192
  "gpt-4o": 128e3,
193
+ "chatgpt-4o-latest": 128e3,
194
+ "gpt-4o-2024-08-06": 128e3,
176
195
  "gpt-4o-2024-05-13": 128e3,
196
+ "gpt-4o-mini": 128e3,
197
+ "gpt-4o-mini-2024-07-18": 128e3,
177
198
  "gpt-4-turbo": 128e3,
178
199
  "gpt-4-turbo-2024-04-09": 128e3,
179
200
  "gpt-4-0125-preview": 128e3,
@@ -219,15 +240,22 @@ function convertActionInputToOpenAITool(action) {
219
240
  function: {
220
241
  name: action.name,
221
242
  description: action.description,
222
- parameters: JSON.parse(action.jsonSchema)
243
+ parameters: (0, import_shared.parseJson)(action.jsonSchema, {})
223
244
  }
224
245
  };
225
246
  }
226
247
  __name(convertActionInputToOpenAITool, "convertActionInputToOpenAITool");
227
- function convertMessageToOpenAIMessage(message) {
248
+ function convertMessageToOpenAIMessage(message, options) {
249
+ const { keepSystemRole } = options || {
250
+ keepSystemRole: false
251
+ };
228
252
  if (message.isTextMessage()) {
253
+ let role = message.role;
254
+ if (message.role === "system" && !keepSystemRole) {
255
+ role = "developer";
256
+ }
229
257
  return {
230
- role: message.role,
258
+ role,
231
259
  content: message.content
232
260
  };
233
261
  } else if (message.isActionExecutionMessage()) {
@@ -256,7 +284,10 @@ __name(convertMessageToOpenAIMessage, "convertMessageToOpenAIMessage");
256
284
  function convertSystemMessageToAssistantAPI(message) {
257
285
  return {
258
286
  ...message,
259
- ...message.role === "system" && {
287
+ ...[
288
+ "system",
289
+ "developer"
290
+ ].includes(message.role) && {
260
291
  role: "assistant",
261
292
  content: "THE FOLLOWING MESSAGE IS A SYSTEM MESSAGE: " + message.content
262
293
  }
@@ -265,7 +296,7 @@ function convertSystemMessageToAssistantAPI(message) {
265
296
  __name(convertSystemMessageToAssistantAPI, "convertSystemMessageToAssistantAPI");
266
297
 
267
298
  // src/service-adapters/openai/openai-adapter.ts
268
- var import_shared = require("@copilotkit/shared");
299
+ var import_shared2 = require("@copilotkit/shared");
269
300
  var DEFAULT_MODEL = "gpt-4o";
270
301
  var OpenAIAdapter = class {
271
302
  model = DEFAULT_MODEL;
@@ -284,8 +315,8 @@ var OpenAIAdapter = class {
284
315
  async process(request) {
285
316
  const { threadId: threadIdFromRequest, model = this.model, messages, actions, eventSource, forwardedParameters } = request;
286
317
  const tools = actions.map(convertActionInputToOpenAITool);
287
- const threadId = threadIdFromRequest ?? (0, import_shared.randomUUID)();
288
- let openaiMessages = messages.map(convertMessageToOpenAIMessage);
318
+ const threadId = threadIdFromRequest ?? (0, import_shared2.randomUUID)();
319
+ let openaiMessages = messages.map((m) => convertMessageToOpenAIMessage(m));
289
320
  openaiMessages = limitMessagesToTokenCount(openaiMessages, tools, model);
290
321
  let toolChoice = forwardedParameters == null ? void 0 : forwardedParameters.toolChoice;
291
322
  if ((forwardedParameters == null ? void 0 : forwardedParameters.toolChoice) === "function") {
@@ -391,7 +422,7 @@ __name(OpenAIAdapter, "OpenAIAdapter");
391
422
  // src/service-adapters/langchain/utils.ts
392
423
  var import_messages = require("@langchain/core/messages");
393
424
  var import_tools = require("@langchain/core/tools");
394
- var import_shared2 = require("@copilotkit/shared");
425
+ var import_shared3 = require("@copilotkit/shared");
395
426
  function convertMessageToLangChainMessage(message) {
396
427
  if (message.isTextMessage()) {
397
428
  if (message.role == "user") {
@@ -424,7 +455,7 @@ function convertActionInputToLangChainTool(actionInput) {
424
455
  return new import_tools.DynamicStructuredTool({
425
456
  name: actionInput.name,
426
457
  description: actionInput.description,
427
- schema: (0, import_shared2.convertJsonSchemaToZodSchema)(JSON.parse(actionInput.jsonSchema), true),
458
+ schema: (0, import_shared3.convertJsonSchemaToZodSchema)(JSON.parse(actionInput.jsonSchema), true),
428
459
  func: async () => {
429
460
  return "";
430
461
  }
@@ -457,7 +488,7 @@ async function streamLangChainResponse({ result, eventStream$, actionExecution }
457
488
  var _a, _b, _c, _d, _e, _f, _g, _h, _i, _j, _k, _l;
458
489
  if (typeof result === "string") {
459
490
  if (!actionExecution) {
460
- eventStream$.sendTextMessage((0, import_shared2.randomId)(), result);
491
+ eventStream$.sendTextMessage((0, import_shared3.randomId)(), result);
461
492
  } else {
462
493
  eventStream$.sendActionExecutionResult({
463
494
  actionExecutionId: actionExecution.id,
@@ -468,11 +499,11 @@ async function streamLangChainResponse({ result, eventStream$, actionExecution }
468
499
  } else if (isAIMessage(result)) {
469
500
  maybeSendActionExecutionResultIsMessage(eventStream$, actionExecution);
470
501
  if (result.content) {
471
- eventStream$.sendTextMessage((0, import_shared2.randomId)(), result.content);
502
+ eventStream$.sendTextMessage((0, import_shared3.randomId)(), result.content);
472
503
  }
473
504
  for (const toolCall of result.tool_calls) {
474
505
  eventStream$.sendActionExecution({
475
- actionExecutionId: toolCall.id || (0, import_shared2.randomId)(),
506
+ actionExecutionId: toolCall.id || (0, import_shared3.randomId)(),
476
507
  actionName: toolCall.name,
477
508
  args: JSON.stringify(toolCall.args)
478
509
  });
@@ -480,12 +511,12 @@ async function streamLangChainResponse({ result, eventStream$, actionExecution }
480
511
  } else if (isBaseMessageChunk(result)) {
481
512
  maybeSendActionExecutionResultIsMessage(eventStream$, actionExecution);
482
513
  if ((_a = result.lc_kwargs) == null ? void 0 : _a.content) {
483
- eventStream$.sendTextMessage((0, import_shared2.randomId)(), result.content);
514
+ eventStream$.sendTextMessage((0, import_shared3.randomId)(), result.content);
484
515
  }
485
516
  if ((_b = result.lc_kwargs) == null ? void 0 : _b.tool_calls) {
486
517
  for (const toolCall of (_c = result.lc_kwargs) == null ? void 0 : _c.tool_calls) {
487
518
  eventStream$.sendActionExecution({
488
- actionExecutionId: toolCall.id || (0, import_shared2.randomId)(),
519
+ actionExecutionId: toolCall.id || (0, import_shared3.randomId)(),
489
520
  actionName: toolCall.name,
490
521
  args: JSON.stringify(toolCall.args)
491
522
  });
@@ -559,7 +590,7 @@ async function streamLangChainResponse({ result, eventStream$, actionExecution }
559
590
  });
560
591
  } else if (content) {
561
592
  mode = "message";
562
- currentMessageId = ((_k = value.lc_kwargs) == null ? void 0 : _k.id) || (0, import_shared2.randomId)();
593
+ currentMessageId = ((_k = value.lc_kwargs) == null ? void 0 : _k.id) || (0, import_shared3.randomId)();
563
594
  eventStream$.sendTextMessageStart({
564
595
  messageId: currentMessageId
565
596
  });
@@ -616,7 +647,7 @@ function encodeResult(result) {
616
647
  __name(encodeResult, "encodeResult");
617
648
 
618
649
  // src/service-adapters/langchain/langchain-adapter.ts
619
- var import_shared3 = require("@copilotkit/shared");
650
+ var import_shared4 = require("@copilotkit/shared");
620
651
  var import_promises = require("@langchain/core/callbacks/promises");
621
652
  var LangChainAdapter = class {
622
653
  options;
@@ -629,7 +660,7 @@ var LangChainAdapter = class {
629
660
  async process(request) {
630
661
  try {
631
662
  const { eventSource, model, actions, messages, runId, threadId: threadIdFromRequest } = request;
632
- const threadId = threadIdFromRequest ?? (0, import_shared3.randomUUID)();
663
+ const threadId = threadIdFromRequest ?? (0, import_shared4.randomUUID)();
633
664
  const result = await this.options.chainFn({
634
665
  messages: messages.map(convertMessageToLangChainMessage),
635
666
  tools: actions.map(convertActionInputToLangChainTool),
@@ -748,7 +779,7 @@ var OpenAIAssistantAdapter = class {
748
779
  ];
749
780
  const instructionsMessage = messages.shift();
750
781
  const instructions = instructionsMessage.isTextMessage() ? instructionsMessage.content : "";
751
- const userMessage = messages.map(convertMessageToOpenAIMessage).map(convertSystemMessageToAssistantAPI).at(-1);
782
+ const userMessage = messages.map((m) => convertMessageToOpenAIMessage(m)).map(convertSystemMessageToAssistantAPI).at(-1);
752
783
  if (userMessage.role !== "user") {
753
784
  throw new Error("No user message found");
754
785
  }
@@ -873,7 +904,7 @@ __name(getRunIdFromStream, "getRunIdFromStream");
873
904
 
874
905
  // src/service-adapters/unify/unify-adapter.ts
875
906
  var import_openai3 = __toESM(require("openai"));
876
- var import_shared4 = require("@copilotkit/shared");
907
+ var import_shared5 = require("@copilotkit/shared");
877
908
  var UnifyAdapter = class {
878
909
  apiKey;
879
910
  model;
@@ -894,7 +925,7 @@ var UnifyAdapter = class {
894
925
  baseURL: "https://api.unify.ai/v0/"
895
926
  });
896
927
  const forwardedParameters = request.forwardedParameters;
897
- const messages = request.messages.map(convertMessageToOpenAIMessage);
928
+ const messages = request.messages.map((m) => convertMessageToOpenAIMessage(m));
898
929
  const stream = await openai.chat.completions.create({
899
930
  model: this.model,
900
931
  messages,
@@ -915,7 +946,7 @@ var UnifyAdapter = class {
915
946
  for await (const chunk of stream) {
916
947
  if (this.start) {
917
948
  model = chunk.model;
918
- currentMessageId = (0, import_shared4.randomId)();
949
+ currentMessageId = (0, import_shared5.randomId)();
919
950
  eventStream$.sendTextMessageStart({
920
951
  messageId: currentMessageId
921
952
  });
@@ -982,7 +1013,7 @@ var UnifyAdapter = class {
982
1013
  eventStream$.complete();
983
1014
  });
984
1015
  return {
985
- threadId: request.threadId || (0, import_shared4.randomUUID)()
1016
+ threadId: request.threadId || (0, import_shared5.randomUUID)()
986
1017
  };
987
1018
  }
988
1019
  };
@@ -990,8 +1021,8 @@ __name(UnifyAdapter, "UnifyAdapter");
990
1021
 
991
1022
  // src/service-adapters/groq/groq-adapter.ts
992
1023
  var import_groq_sdk = require("groq-sdk");
993
- var import_shared5 = require("@copilotkit/shared");
994
- var DEFAULT_MODEL2 = "llama3-groq-70b-8192-tool-use-preview";
1024
+ var import_shared6 = require("@copilotkit/shared");
1025
+ var DEFAULT_MODEL2 = "llama-3.3-70b-versatile";
995
1026
  var GroqAdapter = class {
996
1027
  model = DEFAULT_MODEL2;
997
1028
  disableParallelToolCalls = false;
@@ -1009,7 +1040,9 @@ var GroqAdapter = class {
1009
1040
  async process(request) {
1010
1041
  const { threadId, model = this.model, messages, actions, eventSource, forwardedParameters } = request;
1011
1042
  const tools = actions.map(convertActionInputToOpenAITool);
1012
- let openaiMessages = messages.map(convertMessageToOpenAIMessage);
1043
+ let openaiMessages = messages.map((m) => convertMessageToOpenAIMessage(m, {
1044
+ keepSystemRole: true
1045
+ }));
1013
1046
  openaiMessages = limitMessagesToTokenCount(openaiMessages, tools, model);
1014
1047
  let toolChoice = forwardedParameters == null ? void 0 : forwardedParameters.toolChoice;
1015
1048
  if ((forwardedParameters == null ? void 0 : forwardedParameters.toolChoice) === "function") {
@@ -1103,7 +1136,7 @@ var GroqAdapter = class {
1103
1136
  eventStream$.complete();
1104
1137
  });
1105
1138
  return {
1106
- threadId: request.threadId || (0, import_shared5.randomUUID)()
1139
+ threadId: request.threadId || (0, import_shared6.randomUUID)()
1107
1140
  };
1108
1141
  }
1109
1142
  };
@@ -1245,7 +1278,7 @@ function groupAnthropicMessagesByRole(messageParams) {
1245
1278
  __name(groupAnthropicMessagesByRole, "groupAnthropicMessagesByRole");
1246
1279
 
1247
1280
  // src/service-adapters/anthropic/anthropic-adapter.ts
1248
- var import_shared6 = require("@copilotkit/shared");
1281
+ var import_shared7 = require("@copilotkit/shared");
1249
1282
  var DEFAULT_MODEL3 = "claude-3-sonnet-20240229";
1250
1283
  var AnthropicAdapter = class {
1251
1284
  model = DEFAULT_MODEL3;
@@ -1296,8 +1329,8 @@ var AnthropicAdapter = class {
1296
1329
  eventSource.stream(async (eventStream$) => {
1297
1330
  let mode = null;
1298
1331
  let didOutputText = false;
1299
- let currentMessageId = (0, import_shared6.randomId)();
1300
- let currentToolCallId = (0, import_shared6.randomId)();
1332
+ let currentMessageId = (0, import_shared7.randomId)();
1333
+ let currentToolCallId = (0, import_shared7.randomId)();
1301
1334
  let filterThinkingTextBuffer = new FilterThinkingTextBuffer();
1302
1335
  for await (const chunk of await stream) {
1303
1336
  if (chunk.type === "message_start") {
@@ -1354,7 +1387,7 @@ var AnthropicAdapter = class {
1354
1387
  eventStream$.complete();
1355
1388
  });
1356
1389
  return {
1357
- threadId: threadId || (0, import_shared6.randomUUID)()
1390
+ threadId: threadId || (0, import_shared7.randomUUID)()
1358
1391
  };
1359
1392
  }
1360
1393
  };
@@ -1394,7 +1427,7 @@ var FilterThinkingTextBuffer = /* @__PURE__ */ __name(class FilterThinkingTextBu
1394
1427
 
1395
1428
  // src/service-adapters/experimental/ollama/ollama-adapter.ts
1396
1429
  var import_ollama = require("@langchain/community/llms/ollama");
1397
- var import_shared7 = require("@copilotkit/shared");
1430
+ var import_shared8 = require("@copilotkit/shared");
1398
1431
  var DEFAULT_MODEL4 = "llama3:latest";
1399
1432
  var ExperimentalOllamaAdapter = class {
1400
1433
  model;
@@ -1413,7 +1446,7 @@ var ExperimentalOllamaAdapter = class {
1413
1446
  const contents = messages.filter((m) => m.isTextMessage()).map((m) => m.content);
1414
1447
  const _stream = await ollama.stream(contents);
1415
1448
  eventSource.stream(async (eventStream$) => {
1416
- const currentMessageId = (0, import_shared7.randomId)();
1449
+ const currentMessageId = (0, import_shared8.randomId)();
1417
1450
  eventStream$.sendTextMessageStart({
1418
1451
  messageId: currentMessageId
1419
1452
  });
@@ -1429,18 +1462,18 @@ var ExperimentalOllamaAdapter = class {
1429
1462
  eventStream$.complete();
1430
1463
  });
1431
1464
  return {
1432
- threadId: request.threadId || (0, import_shared7.randomUUID)()
1465
+ threadId: request.threadId || (0, import_shared8.randomUUID)()
1433
1466
  };
1434
1467
  }
1435
1468
  };
1436
1469
  __name(ExperimentalOllamaAdapter, "ExperimentalOllamaAdapter");
1437
1470
 
1438
1471
  // src/service-adapters/empty/empty-adapter.ts
1439
- var import_shared8 = require("@copilotkit/shared");
1472
+ var import_shared9 = require("@copilotkit/shared");
1440
1473
  var EmptyAdapter = class {
1441
1474
  async process(request) {
1442
1475
  return {
1443
- threadId: request.threadId || (0, import_shared8.randomUUID)()
1476
+ threadId: request.threadId || (0, import_shared9.randomUUID)()
1444
1477
  };
1445
1478
  }
1446
1479
  };