@copilotkit/runtime 1.6.0-next.1 → 1.6.0-next.11
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +75 -0
- package/__snapshots__/schema/schema.graphql +1 -0
- package/dist/{chunk-CCQ73DAH.mjs → chunk-6L2NPMPT.mjs} +2 -2
- package/dist/{chunk-CHDIEE43.mjs → chunk-DVFBTXBT.mjs} +87 -44
- package/dist/chunk-DVFBTXBT.mjs.map +1 -0
- package/dist/{chunk-OS5YD32G.mjs → chunk-FZJAYGIR.mjs} +45 -12
- package/dist/chunk-FZJAYGIR.mjs.map +1 -0
- package/dist/{chunk-C7GTLEVO.mjs → chunk-LGQL6LK4.mjs} +2 -2
- package/dist/{chunk-A25FIW7J.mjs → chunk-O3Q4BRZA.mjs} +2 -2
- package/dist/{copilot-runtime-67033bfa.d.ts → copilot-runtime-15bfc4f4.d.ts} +2 -2
- package/dist/graphql/types/converted/index.d.ts +1 -1
- package/dist/{groq-adapter-9d15c927.d.ts → groq-adapter-fb9aa3ab.d.ts} +1 -1
- package/dist/{index-f6d1f30b.d.ts → index-5bec5424.d.ts} +2 -1
- package/dist/index.d.ts +4 -4
- package/dist/index.js +199 -123
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +5 -5
- package/dist/{langserve-7cc5be48.d.ts → langserve-6f7af8d3.d.ts} +1 -1
- package/dist/lib/index.d.ts +4 -4
- package/dist/lib/index.js +192 -116
- package/dist/lib/index.js.map +1 -1
- package/dist/lib/index.mjs +5 -5
- package/dist/lib/integrations/index.d.ts +4 -4
- package/dist/lib/integrations/index.js +5 -3
- package/dist/lib/integrations/index.js.map +1 -1
- package/dist/lib/integrations/index.mjs +5 -5
- package/dist/lib/integrations/nest/index.d.ts +3 -3
- package/dist/lib/integrations/nest/index.js +5 -3
- package/dist/lib/integrations/nest/index.js.map +1 -1
- package/dist/lib/integrations/nest/index.mjs +3 -3
- package/dist/lib/integrations/node-express/index.d.ts +3 -3
- package/dist/lib/integrations/node-express/index.js +5 -3
- package/dist/lib/integrations/node-express/index.js.map +1 -1
- package/dist/lib/integrations/node-express/index.mjs +3 -3
- package/dist/lib/integrations/node-http/index.d.ts +3 -3
- package/dist/lib/integrations/node-http/index.js +5 -3
- package/dist/lib/integrations/node-http/index.js.map +1 -1
- package/dist/lib/integrations/node-http/index.mjs +2 -2
- package/dist/service-adapters/index.d.ts +4 -4
- package/dist/service-adapters/index.js +70 -37
- package/dist/service-adapters/index.js.map +1 -1
- package/dist/service-adapters/index.mjs +1 -1
- package/package.json +5 -4
- package/src/agents/langgraph/event-source.ts +21 -5
- package/src/graphql/types/enums.ts +1 -0
- package/src/lib/runtime/__tests__/remote-action-constructors.test.ts +236 -0
- package/src/lib/runtime/copilot-runtime.ts +9 -3
- package/src/lib/runtime/remote-action-constructors.ts +9 -7
- package/src/lib/runtime/remote-lg-action.ts +35 -7
- package/src/service-adapters/conversion.ts +39 -46
- package/src/service-adapters/groq/groq-adapter.ts +6 -3
- package/src/service-adapters/openai/openai-adapter.ts +1 -1
- package/src/service-adapters/openai/openai-assistant-adapter.ts +1 -1
- package/src/service-adapters/openai/utils.ts +39 -13
- package/src/service-adapters/unify/unify-adapter.ts +1 -1
- package/tsconfig.json +3 -2
- package/dist/chunk-CHDIEE43.mjs.map +0 -1
- package/dist/chunk-OS5YD32G.mjs.map +0 -1
- /package/dist/{chunk-CCQ73DAH.mjs.map → chunk-6L2NPMPT.mjs.map} +0 -0
- /package/dist/{chunk-C7GTLEVO.mjs.map → chunk-LGQL6LK4.mjs.map} +0 -0
- /package/dist/{chunk-A25FIW7J.mjs.map → chunk-O3Q4BRZA.mjs.map} +0 -0
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
import {
|
|
2
2
|
copilotRuntimeNodeHttpEndpoint
|
|
3
|
-
} from "../../../chunk-
|
|
4
|
-
import "../../../chunk-
|
|
3
|
+
} from "../../../chunk-DVFBTXBT.mjs";
|
|
4
|
+
import "../../../chunk-FZJAYGIR.mjs";
|
|
5
5
|
import "../../../chunk-5BIEM2UU.mjs";
|
|
6
6
|
import "../../../chunk-RTFJTJMA.mjs";
|
|
7
7
|
import "../../../chunk-2OZAGFV3.mjs";
|
|
@@ -1,8 +1,8 @@
|
|
|
1
|
-
import { b as CopilotServiceAdapter, C as CopilotRuntimeChatCompletionRequest, a as CopilotRuntimeChatCompletionResponse } from '../langserve-
|
|
2
|
-
export { c as RemoteChain, R as RemoteChainParameters } from '../langserve-
|
|
3
|
-
export { G as GoogleGenerativeAIAdapter, f as GroqAdapter, e as GroqAdapterParams, L as LangChainAdapter, a as OpenAIAdapter, O as OpenAIAdapterParams, c as OpenAIAssistantAdapter, b as OpenAIAssistantAdapterParams, d as UnifyAdapter, U as UnifyAdapterParams } from '../groq-adapter-
|
|
1
|
+
import { b as CopilotServiceAdapter, C as CopilotRuntimeChatCompletionRequest, a as CopilotRuntimeChatCompletionResponse } from '../langserve-6f7af8d3.js';
|
|
2
|
+
export { c as RemoteChain, R as RemoteChainParameters } from '../langserve-6f7af8d3.js';
|
|
3
|
+
export { G as GoogleGenerativeAIAdapter, f as GroqAdapter, e as GroqAdapterParams, L as LangChainAdapter, a as OpenAIAdapter, O as OpenAIAdapterParams, c as OpenAIAssistantAdapter, b as OpenAIAssistantAdapterParams, d as UnifyAdapter, U as UnifyAdapterParams } from '../groq-adapter-fb9aa3ab.js';
|
|
4
4
|
import Anthropic from '@anthropic-ai/sdk';
|
|
5
|
-
import '../index-
|
|
5
|
+
import '../index-5bec5424.js';
|
|
6
6
|
import '../graphql/types/base/index.js';
|
|
7
7
|
import 'rxjs';
|
|
8
8
|
import '@copilotkit/shared';
|
|
@@ -126,6 +126,7 @@ __name(RemoteChain, "RemoteChain");
|
|
|
126
126
|
var import_openai = __toESM(require("openai"));
|
|
127
127
|
|
|
128
128
|
// src/service-adapters/openai/utils.ts
|
|
129
|
+
var import_shared = require("@copilotkit/shared");
|
|
129
130
|
function limitMessagesToTokenCount(messages, tools, model, maxTokens) {
|
|
130
131
|
maxTokens || (maxTokens = maxTokensForOpenAIModel(model));
|
|
131
132
|
const result = [];
|
|
@@ -135,7 +136,10 @@ function limitMessagesToTokenCount(messages, tools, model, maxTokens) {
|
|
|
135
136
|
}
|
|
136
137
|
maxTokens -= toolsNumTokens;
|
|
137
138
|
for (const message of messages) {
|
|
138
|
-
if (
|
|
139
|
+
if ([
|
|
140
|
+
"system",
|
|
141
|
+
"developer"
|
|
142
|
+
].includes(message.role)) {
|
|
139
143
|
const numTokens = countMessageTokens(model, message);
|
|
140
144
|
maxTokens -= numTokens;
|
|
141
145
|
if (maxTokens < 0) {
|
|
@@ -148,7 +152,10 @@ function limitMessagesToTokenCount(messages, tools, model, maxTokens) {
|
|
|
148
152
|
...messages
|
|
149
153
|
].reverse();
|
|
150
154
|
for (const message of reversedMessages) {
|
|
151
|
-
if (
|
|
155
|
+
if ([
|
|
156
|
+
"system",
|
|
157
|
+
"developer"
|
|
158
|
+
].includes(message.role)) {
|
|
152
159
|
result.unshift(message);
|
|
153
160
|
continue;
|
|
154
161
|
} else if (cutoff) {
|
|
@@ -171,9 +178,23 @@ function maxTokensForOpenAIModel(model) {
|
|
|
171
178
|
__name(maxTokensForOpenAIModel, "maxTokensForOpenAIModel");
|
|
172
179
|
var DEFAULT_MAX_TOKENS = 128e3;
|
|
173
180
|
var maxTokensByModel = {
|
|
181
|
+
// o1
|
|
182
|
+
o1: 2e5,
|
|
183
|
+
"o1-2024-12-17": 2e5,
|
|
184
|
+
"o1-mini": 128e3,
|
|
185
|
+
"o1-mini-2024-09-12": 128e3,
|
|
186
|
+
"o1-preview": 128e3,
|
|
187
|
+
"o1-preview-2024-09-12": 128e3,
|
|
188
|
+
// o3-mini
|
|
189
|
+
"o3-mini": 2e5,
|
|
190
|
+
"o3-mini-2025-01-31": 2e5,
|
|
174
191
|
// GPT-4
|
|
175
192
|
"gpt-4o": 128e3,
|
|
193
|
+
"chatgpt-4o-latest": 128e3,
|
|
194
|
+
"gpt-4o-2024-08-06": 128e3,
|
|
176
195
|
"gpt-4o-2024-05-13": 128e3,
|
|
196
|
+
"gpt-4o-mini": 128e3,
|
|
197
|
+
"gpt-4o-mini-2024-07-18": 128e3,
|
|
177
198
|
"gpt-4-turbo": 128e3,
|
|
178
199
|
"gpt-4-turbo-2024-04-09": 128e3,
|
|
179
200
|
"gpt-4-0125-preview": 128e3,
|
|
@@ -219,15 +240,22 @@ function convertActionInputToOpenAITool(action) {
|
|
|
219
240
|
function: {
|
|
220
241
|
name: action.name,
|
|
221
242
|
description: action.description,
|
|
222
|
-
parameters:
|
|
243
|
+
parameters: (0, import_shared.parseJson)(action.jsonSchema, {})
|
|
223
244
|
}
|
|
224
245
|
};
|
|
225
246
|
}
|
|
226
247
|
__name(convertActionInputToOpenAITool, "convertActionInputToOpenAITool");
|
|
227
|
-
function convertMessageToOpenAIMessage(message) {
|
|
248
|
+
function convertMessageToOpenAIMessage(message, options) {
|
|
249
|
+
const { keepSystemRole } = options || {
|
|
250
|
+
keepSystemRole: false
|
|
251
|
+
};
|
|
228
252
|
if (message.isTextMessage()) {
|
|
253
|
+
let role = message.role;
|
|
254
|
+
if (message.role === "system" && !keepSystemRole) {
|
|
255
|
+
role = "developer";
|
|
256
|
+
}
|
|
229
257
|
return {
|
|
230
|
-
role
|
|
258
|
+
role,
|
|
231
259
|
content: message.content
|
|
232
260
|
};
|
|
233
261
|
} else if (message.isActionExecutionMessage()) {
|
|
@@ -256,7 +284,10 @@ __name(convertMessageToOpenAIMessage, "convertMessageToOpenAIMessage");
|
|
|
256
284
|
function convertSystemMessageToAssistantAPI(message) {
|
|
257
285
|
return {
|
|
258
286
|
...message,
|
|
259
|
-
...
|
|
287
|
+
...[
|
|
288
|
+
"system",
|
|
289
|
+
"developer"
|
|
290
|
+
].includes(message.role) && {
|
|
260
291
|
role: "assistant",
|
|
261
292
|
content: "THE FOLLOWING MESSAGE IS A SYSTEM MESSAGE: " + message.content
|
|
262
293
|
}
|
|
@@ -265,7 +296,7 @@ function convertSystemMessageToAssistantAPI(message) {
|
|
|
265
296
|
__name(convertSystemMessageToAssistantAPI, "convertSystemMessageToAssistantAPI");
|
|
266
297
|
|
|
267
298
|
// src/service-adapters/openai/openai-adapter.ts
|
|
268
|
-
var
|
|
299
|
+
var import_shared2 = require("@copilotkit/shared");
|
|
269
300
|
var DEFAULT_MODEL = "gpt-4o";
|
|
270
301
|
var OpenAIAdapter = class {
|
|
271
302
|
model = DEFAULT_MODEL;
|
|
@@ -284,8 +315,8 @@ var OpenAIAdapter = class {
|
|
|
284
315
|
async process(request) {
|
|
285
316
|
const { threadId: threadIdFromRequest, model = this.model, messages, actions, eventSource, forwardedParameters } = request;
|
|
286
317
|
const tools = actions.map(convertActionInputToOpenAITool);
|
|
287
|
-
const threadId = threadIdFromRequest ?? (0,
|
|
288
|
-
let openaiMessages = messages.map(convertMessageToOpenAIMessage);
|
|
318
|
+
const threadId = threadIdFromRequest ?? (0, import_shared2.randomUUID)();
|
|
319
|
+
let openaiMessages = messages.map((m) => convertMessageToOpenAIMessage(m));
|
|
289
320
|
openaiMessages = limitMessagesToTokenCount(openaiMessages, tools, model);
|
|
290
321
|
let toolChoice = forwardedParameters == null ? void 0 : forwardedParameters.toolChoice;
|
|
291
322
|
if ((forwardedParameters == null ? void 0 : forwardedParameters.toolChoice) === "function") {
|
|
@@ -391,7 +422,7 @@ __name(OpenAIAdapter, "OpenAIAdapter");
|
|
|
391
422
|
// src/service-adapters/langchain/utils.ts
|
|
392
423
|
var import_messages = require("@langchain/core/messages");
|
|
393
424
|
var import_tools = require("@langchain/core/tools");
|
|
394
|
-
var
|
|
425
|
+
var import_shared3 = require("@copilotkit/shared");
|
|
395
426
|
function convertMessageToLangChainMessage(message) {
|
|
396
427
|
if (message.isTextMessage()) {
|
|
397
428
|
if (message.role == "user") {
|
|
@@ -424,7 +455,7 @@ function convertActionInputToLangChainTool(actionInput) {
|
|
|
424
455
|
return new import_tools.DynamicStructuredTool({
|
|
425
456
|
name: actionInput.name,
|
|
426
457
|
description: actionInput.description,
|
|
427
|
-
schema: (0,
|
|
458
|
+
schema: (0, import_shared3.convertJsonSchemaToZodSchema)(JSON.parse(actionInput.jsonSchema), true),
|
|
428
459
|
func: async () => {
|
|
429
460
|
return "";
|
|
430
461
|
}
|
|
@@ -457,7 +488,7 @@ async function streamLangChainResponse({ result, eventStream$, actionExecution }
|
|
|
457
488
|
var _a, _b, _c, _d, _e, _f, _g, _h, _i, _j, _k, _l;
|
|
458
489
|
if (typeof result === "string") {
|
|
459
490
|
if (!actionExecution) {
|
|
460
|
-
eventStream$.sendTextMessage((0,
|
|
491
|
+
eventStream$.sendTextMessage((0, import_shared3.randomId)(), result);
|
|
461
492
|
} else {
|
|
462
493
|
eventStream$.sendActionExecutionResult({
|
|
463
494
|
actionExecutionId: actionExecution.id,
|
|
@@ -468,11 +499,11 @@ async function streamLangChainResponse({ result, eventStream$, actionExecution }
|
|
|
468
499
|
} else if (isAIMessage(result)) {
|
|
469
500
|
maybeSendActionExecutionResultIsMessage(eventStream$, actionExecution);
|
|
470
501
|
if (result.content) {
|
|
471
|
-
eventStream$.sendTextMessage((0,
|
|
502
|
+
eventStream$.sendTextMessage((0, import_shared3.randomId)(), result.content);
|
|
472
503
|
}
|
|
473
504
|
for (const toolCall of result.tool_calls) {
|
|
474
505
|
eventStream$.sendActionExecution({
|
|
475
|
-
actionExecutionId: toolCall.id || (0,
|
|
506
|
+
actionExecutionId: toolCall.id || (0, import_shared3.randomId)(),
|
|
476
507
|
actionName: toolCall.name,
|
|
477
508
|
args: JSON.stringify(toolCall.args)
|
|
478
509
|
});
|
|
@@ -480,12 +511,12 @@ async function streamLangChainResponse({ result, eventStream$, actionExecution }
|
|
|
480
511
|
} else if (isBaseMessageChunk(result)) {
|
|
481
512
|
maybeSendActionExecutionResultIsMessage(eventStream$, actionExecution);
|
|
482
513
|
if ((_a = result.lc_kwargs) == null ? void 0 : _a.content) {
|
|
483
|
-
eventStream$.sendTextMessage((0,
|
|
514
|
+
eventStream$.sendTextMessage((0, import_shared3.randomId)(), result.content);
|
|
484
515
|
}
|
|
485
516
|
if ((_b = result.lc_kwargs) == null ? void 0 : _b.tool_calls) {
|
|
486
517
|
for (const toolCall of (_c = result.lc_kwargs) == null ? void 0 : _c.tool_calls) {
|
|
487
518
|
eventStream$.sendActionExecution({
|
|
488
|
-
actionExecutionId: toolCall.id || (0,
|
|
519
|
+
actionExecutionId: toolCall.id || (0, import_shared3.randomId)(),
|
|
489
520
|
actionName: toolCall.name,
|
|
490
521
|
args: JSON.stringify(toolCall.args)
|
|
491
522
|
});
|
|
@@ -559,7 +590,7 @@ async function streamLangChainResponse({ result, eventStream$, actionExecution }
|
|
|
559
590
|
});
|
|
560
591
|
} else if (content) {
|
|
561
592
|
mode = "message";
|
|
562
|
-
currentMessageId = ((_k = value.lc_kwargs) == null ? void 0 : _k.id) || (0,
|
|
593
|
+
currentMessageId = ((_k = value.lc_kwargs) == null ? void 0 : _k.id) || (0, import_shared3.randomId)();
|
|
563
594
|
eventStream$.sendTextMessageStart({
|
|
564
595
|
messageId: currentMessageId
|
|
565
596
|
});
|
|
@@ -616,7 +647,7 @@ function encodeResult(result) {
|
|
|
616
647
|
__name(encodeResult, "encodeResult");
|
|
617
648
|
|
|
618
649
|
// src/service-adapters/langchain/langchain-adapter.ts
|
|
619
|
-
var
|
|
650
|
+
var import_shared4 = require("@copilotkit/shared");
|
|
620
651
|
var import_promises = require("@langchain/core/callbacks/promises");
|
|
621
652
|
var LangChainAdapter = class {
|
|
622
653
|
options;
|
|
@@ -629,7 +660,7 @@ var LangChainAdapter = class {
|
|
|
629
660
|
async process(request) {
|
|
630
661
|
try {
|
|
631
662
|
const { eventSource, model, actions, messages, runId, threadId: threadIdFromRequest } = request;
|
|
632
|
-
const threadId = threadIdFromRequest ?? (0,
|
|
663
|
+
const threadId = threadIdFromRequest ?? (0, import_shared4.randomUUID)();
|
|
633
664
|
const result = await this.options.chainFn({
|
|
634
665
|
messages: messages.map(convertMessageToLangChainMessage),
|
|
635
666
|
tools: actions.map(convertActionInputToLangChainTool),
|
|
@@ -748,7 +779,7 @@ var OpenAIAssistantAdapter = class {
|
|
|
748
779
|
];
|
|
749
780
|
const instructionsMessage = messages.shift();
|
|
750
781
|
const instructions = instructionsMessage.isTextMessage() ? instructionsMessage.content : "";
|
|
751
|
-
const userMessage = messages.map(convertMessageToOpenAIMessage).map(convertSystemMessageToAssistantAPI).at(-1);
|
|
782
|
+
const userMessage = messages.map((m) => convertMessageToOpenAIMessage(m)).map(convertSystemMessageToAssistantAPI).at(-1);
|
|
752
783
|
if (userMessage.role !== "user") {
|
|
753
784
|
throw new Error("No user message found");
|
|
754
785
|
}
|
|
@@ -873,7 +904,7 @@ __name(getRunIdFromStream, "getRunIdFromStream");
|
|
|
873
904
|
|
|
874
905
|
// src/service-adapters/unify/unify-adapter.ts
|
|
875
906
|
var import_openai3 = __toESM(require("openai"));
|
|
876
|
-
var
|
|
907
|
+
var import_shared5 = require("@copilotkit/shared");
|
|
877
908
|
var UnifyAdapter = class {
|
|
878
909
|
apiKey;
|
|
879
910
|
model;
|
|
@@ -894,7 +925,7 @@ var UnifyAdapter = class {
|
|
|
894
925
|
baseURL: "https://api.unify.ai/v0/"
|
|
895
926
|
});
|
|
896
927
|
const forwardedParameters = request.forwardedParameters;
|
|
897
|
-
const messages = request.messages.map(convertMessageToOpenAIMessage);
|
|
928
|
+
const messages = request.messages.map((m) => convertMessageToOpenAIMessage(m));
|
|
898
929
|
const stream = await openai.chat.completions.create({
|
|
899
930
|
model: this.model,
|
|
900
931
|
messages,
|
|
@@ -915,7 +946,7 @@ var UnifyAdapter = class {
|
|
|
915
946
|
for await (const chunk of stream) {
|
|
916
947
|
if (this.start) {
|
|
917
948
|
model = chunk.model;
|
|
918
|
-
currentMessageId = (0,
|
|
949
|
+
currentMessageId = (0, import_shared5.randomId)();
|
|
919
950
|
eventStream$.sendTextMessageStart({
|
|
920
951
|
messageId: currentMessageId
|
|
921
952
|
});
|
|
@@ -982,7 +1013,7 @@ var UnifyAdapter = class {
|
|
|
982
1013
|
eventStream$.complete();
|
|
983
1014
|
});
|
|
984
1015
|
return {
|
|
985
|
-
threadId: request.threadId || (0,
|
|
1016
|
+
threadId: request.threadId || (0, import_shared5.randomUUID)()
|
|
986
1017
|
};
|
|
987
1018
|
}
|
|
988
1019
|
};
|
|
@@ -990,8 +1021,8 @@ __name(UnifyAdapter, "UnifyAdapter");
|
|
|
990
1021
|
|
|
991
1022
|
// src/service-adapters/groq/groq-adapter.ts
|
|
992
1023
|
var import_groq_sdk = require("groq-sdk");
|
|
993
|
-
var
|
|
994
|
-
var DEFAULT_MODEL2 = "
|
|
1024
|
+
var import_shared6 = require("@copilotkit/shared");
|
|
1025
|
+
var DEFAULT_MODEL2 = "llama-3.3-70b-versatile";
|
|
995
1026
|
var GroqAdapter = class {
|
|
996
1027
|
model = DEFAULT_MODEL2;
|
|
997
1028
|
disableParallelToolCalls = false;
|
|
@@ -1009,7 +1040,9 @@ var GroqAdapter = class {
|
|
|
1009
1040
|
async process(request) {
|
|
1010
1041
|
const { threadId, model = this.model, messages, actions, eventSource, forwardedParameters } = request;
|
|
1011
1042
|
const tools = actions.map(convertActionInputToOpenAITool);
|
|
1012
|
-
let openaiMessages = messages.map(convertMessageToOpenAIMessage
|
|
1043
|
+
let openaiMessages = messages.map((m) => convertMessageToOpenAIMessage(m, {
|
|
1044
|
+
keepSystemRole: true
|
|
1045
|
+
}));
|
|
1013
1046
|
openaiMessages = limitMessagesToTokenCount(openaiMessages, tools, model);
|
|
1014
1047
|
let toolChoice = forwardedParameters == null ? void 0 : forwardedParameters.toolChoice;
|
|
1015
1048
|
if ((forwardedParameters == null ? void 0 : forwardedParameters.toolChoice) === "function") {
|
|
@@ -1103,7 +1136,7 @@ var GroqAdapter = class {
|
|
|
1103
1136
|
eventStream$.complete();
|
|
1104
1137
|
});
|
|
1105
1138
|
return {
|
|
1106
|
-
threadId: request.threadId || (0,
|
|
1139
|
+
threadId: request.threadId || (0, import_shared6.randomUUID)()
|
|
1107
1140
|
};
|
|
1108
1141
|
}
|
|
1109
1142
|
};
|
|
@@ -1245,7 +1278,7 @@ function groupAnthropicMessagesByRole(messageParams) {
|
|
|
1245
1278
|
__name(groupAnthropicMessagesByRole, "groupAnthropicMessagesByRole");
|
|
1246
1279
|
|
|
1247
1280
|
// src/service-adapters/anthropic/anthropic-adapter.ts
|
|
1248
|
-
var
|
|
1281
|
+
var import_shared7 = require("@copilotkit/shared");
|
|
1249
1282
|
var DEFAULT_MODEL3 = "claude-3-sonnet-20240229";
|
|
1250
1283
|
var AnthropicAdapter = class {
|
|
1251
1284
|
model = DEFAULT_MODEL3;
|
|
@@ -1296,8 +1329,8 @@ var AnthropicAdapter = class {
|
|
|
1296
1329
|
eventSource.stream(async (eventStream$) => {
|
|
1297
1330
|
let mode = null;
|
|
1298
1331
|
let didOutputText = false;
|
|
1299
|
-
let currentMessageId = (0,
|
|
1300
|
-
let currentToolCallId = (0,
|
|
1332
|
+
let currentMessageId = (0, import_shared7.randomId)();
|
|
1333
|
+
let currentToolCallId = (0, import_shared7.randomId)();
|
|
1301
1334
|
let filterThinkingTextBuffer = new FilterThinkingTextBuffer();
|
|
1302
1335
|
for await (const chunk of await stream) {
|
|
1303
1336
|
if (chunk.type === "message_start") {
|
|
@@ -1354,7 +1387,7 @@ var AnthropicAdapter = class {
|
|
|
1354
1387
|
eventStream$.complete();
|
|
1355
1388
|
});
|
|
1356
1389
|
return {
|
|
1357
|
-
threadId: threadId || (0,
|
|
1390
|
+
threadId: threadId || (0, import_shared7.randomUUID)()
|
|
1358
1391
|
};
|
|
1359
1392
|
}
|
|
1360
1393
|
};
|
|
@@ -1394,7 +1427,7 @@ var FilterThinkingTextBuffer = /* @__PURE__ */ __name(class FilterThinkingTextBu
|
|
|
1394
1427
|
|
|
1395
1428
|
// src/service-adapters/experimental/ollama/ollama-adapter.ts
|
|
1396
1429
|
var import_ollama = require("@langchain/community/llms/ollama");
|
|
1397
|
-
var
|
|
1430
|
+
var import_shared8 = require("@copilotkit/shared");
|
|
1398
1431
|
var DEFAULT_MODEL4 = "llama3:latest";
|
|
1399
1432
|
var ExperimentalOllamaAdapter = class {
|
|
1400
1433
|
model;
|
|
@@ -1413,7 +1446,7 @@ var ExperimentalOllamaAdapter = class {
|
|
|
1413
1446
|
const contents = messages.filter((m) => m.isTextMessage()).map((m) => m.content);
|
|
1414
1447
|
const _stream = await ollama.stream(contents);
|
|
1415
1448
|
eventSource.stream(async (eventStream$) => {
|
|
1416
|
-
const currentMessageId = (0,
|
|
1449
|
+
const currentMessageId = (0, import_shared8.randomId)();
|
|
1417
1450
|
eventStream$.sendTextMessageStart({
|
|
1418
1451
|
messageId: currentMessageId
|
|
1419
1452
|
});
|
|
@@ -1429,18 +1462,18 @@ var ExperimentalOllamaAdapter = class {
|
|
|
1429
1462
|
eventStream$.complete();
|
|
1430
1463
|
});
|
|
1431
1464
|
return {
|
|
1432
|
-
threadId: request.threadId || (0,
|
|
1465
|
+
threadId: request.threadId || (0, import_shared8.randomUUID)()
|
|
1433
1466
|
};
|
|
1434
1467
|
}
|
|
1435
1468
|
};
|
|
1436
1469
|
__name(ExperimentalOllamaAdapter, "ExperimentalOllamaAdapter");
|
|
1437
1470
|
|
|
1438
1471
|
// src/service-adapters/empty/empty-adapter.ts
|
|
1439
|
-
var
|
|
1472
|
+
var import_shared9 = require("@copilotkit/shared");
|
|
1440
1473
|
var EmptyAdapter = class {
|
|
1441
1474
|
async process(request) {
|
|
1442
1475
|
return {
|
|
1443
|
-
threadId: request.threadId || (0,
|
|
1476
|
+
threadId: request.threadId || (0, import_shared9.randomUUID)()
|
|
1444
1477
|
};
|
|
1445
1478
|
}
|
|
1446
1479
|
};
|