@copilotkit/react-core 0.16.0 → 0.17.0-alpha.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.turbo/turbo-build.log +168 -134
- package/CHANGELOG.md +12 -0
- package/dist/chunk-5UGLWBZJ.mjs +3 -0
- package/dist/{chunk-YGJFU4ZP.mjs → chunk-6A4PCNMZ.mjs} +2 -12
- package/dist/chunk-6A4PCNMZ.mjs.map +1 -0
- package/dist/chunk-6SOQYBNX.mjs +44 -0
- package/dist/chunk-6SOQYBNX.mjs.map +1 -0
- package/dist/{chunk-RFZQHCNS.mjs → chunk-7P4OYNP3.mjs} +2 -2
- package/dist/{chunk-WL2MC3E2.mjs → chunk-FP4EXCGS.mjs} +2 -2
- package/dist/{chunk-PF7LXYPO.mjs → chunk-HW5IH6PV.mjs} +2 -2
- package/dist/chunk-PUFR5OAI.mjs +3 -0
- package/dist/{chunk-EV26IMLL.mjs → chunk-QTET3FBV.mjs} +5 -5
- package/dist/chunk-QTET3FBV.mjs.map +1 -0
- package/dist/{chunk-DE37LEZJ.mjs → chunk-SM3XH2NJ.mjs} +10 -12
- package/dist/chunk-SM3XH2NJ.mjs.map +1 -0
- package/dist/chunk-TDH7XE4X.mjs +104 -0
- package/dist/chunk-TDH7XE4X.mjs.map +1 -0
- package/dist/chunk-TKXPMMND.mjs +177 -0
- package/dist/chunk-TKXPMMND.mjs.map +1 -0
- package/dist/{chunk-IF64NU27.mjs → chunk-TU4DNDO7.mjs} +4 -4
- package/dist/chunk-TU4DNDO7.mjs.map +1 -0
- package/dist/chunk-VK25FIML.mjs +144 -0
- package/dist/chunk-VK25FIML.mjs.map +1 -0
- package/dist/chunk-VNRDQJXW.mjs +3 -0
- package/dist/chunk-VNRDQJXW.mjs.map +1 -0
- package/dist/components/copilot-provider/copilot-provider-props.d.ts +3 -3
- package/dist/components/copilot-provider/copilot-provider.d.ts +3 -3
- package/dist/components/copilot-provider/copilot-provider.mjs +2 -2
- package/dist/components/copilot-provider/index.d.ts +3 -3
- package/dist/components/copilot-provider/index.mjs +2 -2
- package/dist/components/copilot-provider/standard-copilot-api-config.d.ts +3 -3
- package/dist/components/index.d.ts +3 -3
- package/dist/components/index.mjs +2 -2
- package/dist/context/copilot-context.d.ts +4 -8
- package/dist/context/copilot-context.mjs +1 -1
- package/dist/context/index.d.ts +4 -4
- package/dist/context/index.mjs +2 -2
- package/dist/hooks/index.d.ts +6 -1
- package/dist/hooks/index.mjs +10 -7
- package/dist/hooks/use-chat.d.ts +84 -0
- package/dist/hooks/use-chat.mjs +6 -0
- package/dist/hooks/use-chat.mjs.map +1 -0
- package/dist/hooks/use-copilot-chat.d.ts +10 -3
- package/dist/hooks/use-copilot-chat.mjs +6 -3
- package/dist/hooks/use-make-copilot-actionable.mjs +2 -2
- package/dist/hooks/use-make-copilot-document-readable.mjs +2 -2
- package/dist/hooks/use-make-copilot-readable.mjs +2 -2
- package/dist/index.d.ts +6 -2
- package/dist/index.mjs +15 -10
- package/dist/openai/chat-completion-client.d.ts +52 -0
- package/dist/openai/chat-completion-client.mjs +5 -0
- package/dist/openai/chat-completion-client.mjs.map +1 -0
- package/dist/openai/chat-completion-stream.d.ts +21 -0
- package/dist/openai/chat-completion-stream.mjs +5 -0
- package/dist/openai/chat-completion-stream.mjs.map +1 -0
- package/dist/openai/chat-completion-transport.d.ts +38 -0
- package/dist/openai/chat-completion-transport.mjs +4 -0
- package/dist/openai/chat-completion-transport.mjs.map +1 -0
- package/dist/openai/index.d.ts +10 -0
- package/dist/openai/index.mjs +6 -0
- package/dist/openai/index.mjs.map +1 -0
- package/dist/openai-assistants/hooks/index.mjs +4 -4
- package/dist/openai-assistants/hooks/use-copilot-chat-v2.mjs +4 -4
- package/dist/openai-assistants/index.mjs +4 -4
- package/dist/types/base.d.ts +56 -0
- package/dist/types/base.mjs +3 -0
- package/dist/types/base.mjs.map +1 -0
- package/dist/types/index.d.ts +1 -0
- package/dist/types/message.d.ts +2 -0
- package/dist/types/message.mjs +3 -0
- package/dist/types/message.mjs.map +1 -0
- package/package.json +2 -2
- package/src/components/copilot-provider/copilot-provider.tsx +4 -5
- package/src/context/copilot-context.tsx +1 -12
- package/src/context/index.ts +0 -1
- package/src/hooks/index.ts +1 -0
- package/src/hooks/use-chat.ts +197 -0
- package/src/hooks/use-copilot-chat.ts +10 -22
- package/src/index.tsx +1 -0
- package/src/openai/chat-completion-client.ts +229 -0
- package/src/openai/chat-completion-stream.ts +54 -0
- package/src/openai/chat-completion-transport.ts +186 -0
- package/src/openai/index.tsx +5 -0
- package/src/openai-assistants/hooks/use-copilot-chat-v2.ts +2 -2
- package/src/types/base.ts +61 -0
- package/src/types/index.ts +1 -0
- package/src/types/message.ts +0 -0
- package/dist/chunk-DE37LEZJ.mjs.map +0 -1
- package/dist/chunk-EV26IMLL.mjs.map +0 -1
- package/dist/chunk-IF64NU27.mjs.map +0 -1
- package/dist/chunk-QACD2U6P.mjs +0 -3
- package/dist/chunk-YGJFU4ZP.mjs.map +0 -1
- package/dist/chunk-YPSGKPDA.mjs +0 -3
- /package/dist/{chunk-QACD2U6P.mjs.map → chunk-5UGLWBZJ.mjs.map} +0 -0
- /package/dist/{chunk-RFZQHCNS.mjs.map → chunk-7P4OYNP3.mjs.map} +0 -0
- /package/dist/{chunk-WL2MC3E2.mjs.map → chunk-FP4EXCGS.mjs.map} +0 -0
- /package/dist/{chunk-PF7LXYPO.mjs.map → chunk-HW5IH6PV.mjs.map} +0 -0
- /package/dist/{chunk-YPSGKPDA.mjs.map → chunk-PUFR5OAI.mjs.map} +0 -0
|
@@ -1,8 +1,8 @@
|
|
|
1
|
-
import { defaultCopilotContextCategories } from './chunk-
|
|
2
|
-
import { CopilotContext
|
|
1
|
+
import { defaultCopilotContextCategories } from './chunk-QTET3FBV.mjs';
|
|
2
|
+
import { CopilotContext } from './chunk-6A4PCNMZ.mjs';
|
|
3
|
+
import { useChat } from './chunk-TDH7XE4X.mjs';
|
|
3
4
|
import { __objRest, __spreadProps, __spreadValues } from './chunk-MRXNTQOX.mjs';
|
|
4
5
|
import { useContext, useMemo } from 'react';
|
|
5
|
-
import { useChat } from 'ai/react';
|
|
6
6
|
|
|
7
7
|
function useCopilotChat(_a) {
|
|
8
8
|
var _b = _a, {
|
|
@@ -25,19 +25,17 @@ function useCopilotChat(_a) {
|
|
|
25
25
|
role: "system"
|
|
26
26
|
};
|
|
27
27
|
}, [getContextString, makeSystemMessage]);
|
|
28
|
-
const initialMessagesWithContext = [systemMessage].concat(options.initialMessages || []);
|
|
29
28
|
const functionDescriptions = useMemo(() => {
|
|
30
29
|
return getChatCompletionFunctionDescriptions();
|
|
31
30
|
}, [getChatCompletionFunctionDescriptions]);
|
|
32
31
|
const { messages, append, reload, stop, isLoading, input, setInput } = useChat(__spreadProps(__spreadValues({}, options), {
|
|
33
|
-
|
|
32
|
+
copilotConfig: copilotApiConfig,
|
|
34
33
|
id: options.id,
|
|
35
|
-
initialMessages:
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
}, functionDescriptions.length > 0 && { functions: functionDescriptions }), copilotApiConfig.body), options.body)
|
|
34
|
+
initialMessages: [systemMessage].concat(options.initialMessages || []),
|
|
35
|
+
functions: functionDescriptions,
|
|
36
|
+
onFunctionCall: getFunctionCallHandler(),
|
|
37
|
+
headers: __spreadValues({}, options.headers),
|
|
38
|
+
body: __spreadValues({}, options.body)
|
|
41
39
|
}));
|
|
42
40
|
const visibleMessages = messages.filter(
|
|
43
41
|
(message) => message.role === "user" || message.role === "assistant"
|
|
@@ -76,4 +74,4 @@ If you would like to call a function, call it without saying anything else.
|
|
|
76
74
|
|
|
77
75
|
export { defaultSystemMessage, useCopilotChat };
|
|
78
76
|
//# sourceMappingURL=out.js.map
|
|
79
|
-
//# sourceMappingURL=chunk-
|
|
77
|
+
//# sourceMappingURL=chunk-SM3XH2NJ.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/hooks/use-copilot-chat.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;AAAA,SAAS,SAAS,kBAAkB;AAqB7B,SAAS,eAAe,IAGiB;AAHjB,eAC7B;AAAA;AAAA,EAtBF,IAqB+B,IAE1B,oBAF0B,IAE1B;AAAA,IADH;AAAA;AAGA,QAAM;AAAA,IACJ;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF,IAAI,WAAW,cAAc;AAE7B,QAAM,gBAAyB,QAAQ,MAAM;AAC3C,UAAM,qBAAqB,qBAAqB;AAChD,UAAM,gBAAgB,iBAAiB,CAAC,GAAG,+BAA+B;AAE1E,WAAO;AAAA,MACL,IAAI;AAAA,MACJ,SAAS,mBAAmB,aAAa;AAAA,MACzC,MAAM;AAAA,IACR;AAAA,EACF,GAAG,CAAC,kBAAkB,iBAAiB,CAAC;AAExC,QAAM,uBAA8D,QAAQ,MAAM;AAChF,WAAO,sCAAsC;AAAA,EAC/C,GAAG,CAAC,qCAAqC,CAAC;AAE1C,QAAM,EAAE,UAAU,QAAQ,QAAQ,MAAM,WAAW,OAAO,SAAS,IAAI,QAAQ,iCAC1E,UAD0E;AAAA,IAE7E,eAAe;AAAA,IACf,IAAI,QAAQ;AAAA,IACZ,iBAAiB,CAAC,aAAa,EAAE,OAAO,QAAQ,mBAAmB,CAAC,CAAC;AAAA,IACrE,WAAW;AAAA,IACX,gBAAgB,uBAAuB;AAAA,IACvC,SAAS,mBAAK,QAAQ;AAAA,IACtB,MAAM,mBACD,QAAQ;AAAA,EAEf,EAAC;AAED,QAAM,kBAAkB,SAAS;AAAA,IAC/B,CAAC,YAAY,QAAQ,SAAS,UAAU,QAAQ,SAAS;AAAA,EAC3D;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;AAEO,SAAS,qBAAqB,eAA+B;AAClE,SAAO;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAQP;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAWF","sourcesContent":["import { useMemo, useContext } from \"react\";\nimport { CopilotContext, CopilotContextParams } from \"../context/copilot-context\";\nimport { Message } from \"../types\";\nimport { UseChatOptions, useChat } from \"./use-chat\";\nimport { defaultCopilotContextCategories } from \"../components\";\nimport { ChatCompletionCreateParams } from \"openai/resources/chat\";\n\nexport interface UseCopilotChatOptions extends UseChatOptions {\n makeSystemMessage?: (contextString: string) => string;\n}\n\nexport interface UseCopilotChatReturn {\n visibleMessages: Message[];\n append: (message: Message) => Promise<void>;\n reload: () => Promise<void>;\n stop: () => void;\n isLoading: boolean;\n input: string;\n setInput: React.Dispatch<React.SetStateAction<string>>;\n}\n\nexport function useCopilotChat({\n makeSystemMessage,\n ...options\n}: UseCopilotChatOptions): UseCopilotChatReturn {\n const {\n getContextString,\n getChatCompletionFunctionDescriptions,\n getFunctionCallHandler,\n copilotApiConfig,\n } = useContext(CopilotContext);\n\n const systemMessage: Message = useMemo(() => {\n const systemMessageMaker = makeSystemMessage || defaultSystemMessage;\n const contextString = getContextString([], defaultCopilotContextCategories); // TODO: make the context categories configurable\n\n return {\n id: \"system\",\n content: systemMessageMaker(contextString),\n role: \"system\",\n };\n }, [getContextString, makeSystemMessage]);\n\n const functionDescriptions: ChatCompletionCreateParams.Function[] = useMemo(() => {\n return getChatCompletionFunctionDescriptions();\n }, [getChatCompletionFunctionDescriptions]);\n\n const { messages, append, reload, stop, isLoading, input, setInput } = useChat({\n ...options,\n copilotConfig: copilotApiConfig,\n id: options.id,\n initialMessages: [systemMessage].concat(options.initialMessages || []),\n functions: functionDescriptions,\n onFunctionCall: getFunctionCallHandler(),\n headers: { ...options.headers },\n body: {\n ...options.body,\n },\n });\n\n const visibleMessages = messages.filter(\n (message) => message.role === \"user\" || message.role === \"assistant\",\n );\n\n return {\n visibleMessages,\n append,\n reload,\n stop,\n isLoading,\n input,\n setInput,\n };\n}\n\nexport function defaultSystemMessage(contextString: string): string {\n return `\nPlease act as an efficient, competent, conscientious, and industrious professional assistant.\n\nHelp the user achieve their goals, and you do so in a way that is as efficient as possible, without unnecessary fluff, but also without sacrificing professionalism.\nAlways be polite and respectful, and prefer brevity over verbosity.\n\nThe user has provided you with the following context:\n\\`\\`\\`\n${contextString}\n\\`\\`\\`\n\nThey have also provided you with functions you can call to initiate actions on their behalf, or functions you can call to receive more information.\n\nPlease assist them as best you can.\n\nYou can ask them for clarifying questions if needed, but don't be annoying about it. If you can reasonably 'fill in the blanks' yourself, do so.\n\nIf you would like to call a function, call it without saying anything else.\n`;\n}\n"]}
|
|
@@ -0,0 +1,104 @@
|
|
|
1
|
+
import { ChatCompletionClient } from './chunk-TKXPMMND.mjs';
|
|
2
|
+
import { __async, __spreadValues } from './chunk-MRXNTQOX.mjs';
|
|
3
|
+
import { useState } from 'react';
|
|
4
|
+
import { nanoid } from 'nanoid';
|
|
5
|
+
|
|
6
|
+
function useChat(options) {
|
|
7
|
+
const [messages, setMessages] = useState([]);
|
|
8
|
+
const [input, setInput] = useState("");
|
|
9
|
+
const [isLoading, setIsLoading] = useState(false);
|
|
10
|
+
const runChatCompletion = (messages2) => __async(this, null, function* () {
|
|
11
|
+
return new Promise((resolve, reject) => {
|
|
12
|
+
setIsLoading(true);
|
|
13
|
+
const assistantMessage = {
|
|
14
|
+
id: nanoid(),
|
|
15
|
+
createdAt: /* @__PURE__ */ new Date(),
|
|
16
|
+
content: "",
|
|
17
|
+
role: "assistant"
|
|
18
|
+
};
|
|
19
|
+
setMessages([...messages2, __spreadValues({}, assistantMessage)]);
|
|
20
|
+
const messagesWithContext = [...options.initialMessages || [], ...messages2];
|
|
21
|
+
const client = new ChatCompletionClient({
|
|
22
|
+
url: options.api
|
|
23
|
+
});
|
|
24
|
+
const cleanup = () => {
|
|
25
|
+
client.off("content");
|
|
26
|
+
client.off("end");
|
|
27
|
+
client.off("error");
|
|
28
|
+
client.off("function");
|
|
29
|
+
};
|
|
30
|
+
client.on("content", (content) => {
|
|
31
|
+
assistantMessage.content += content;
|
|
32
|
+
setMessages([...messages2, __spreadValues({}, assistantMessage)]);
|
|
33
|
+
});
|
|
34
|
+
client.on("end", () => {
|
|
35
|
+
setIsLoading(false);
|
|
36
|
+
cleanup();
|
|
37
|
+
resolve(__spreadValues({}, assistantMessage));
|
|
38
|
+
});
|
|
39
|
+
client.on("error", (error) => {
|
|
40
|
+
setIsLoading(false);
|
|
41
|
+
cleanup();
|
|
42
|
+
reject(error);
|
|
43
|
+
});
|
|
44
|
+
client.on("function", (functionCall) => __async(this, null, function* () {
|
|
45
|
+
assistantMessage.function_call = {
|
|
46
|
+
name: functionCall.name,
|
|
47
|
+
arguments: JSON.stringify(functionCall.arguments)
|
|
48
|
+
};
|
|
49
|
+
setMessages([...messages2, __spreadValues({}, assistantMessage)]);
|
|
50
|
+
setIsLoading(false);
|
|
51
|
+
cleanup();
|
|
52
|
+
resolve(__spreadValues({}, assistantMessage));
|
|
53
|
+
}));
|
|
54
|
+
client.fetch({
|
|
55
|
+
copilotConfig: options.copilotConfig,
|
|
56
|
+
messages: messagesWithContext,
|
|
57
|
+
functions: options.functions,
|
|
58
|
+
headers: options.headers
|
|
59
|
+
});
|
|
60
|
+
});
|
|
61
|
+
});
|
|
62
|
+
const runChatCompletionAndHandleFunctionCall = (messages2) => __async(this, null, function* () {
|
|
63
|
+
const message = yield runChatCompletion(messages2);
|
|
64
|
+
if (message.function_call && options.onFunctionCall) {
|
|
65
|
+
yield options.onFunctionCall(messages2, message.function_call);
|
|
66
|
+
}
|
|
67
|
+
});
|
|
68
|
+
const append = (message) => __async(this, null, function* () {
|
|
69
|
+
if (isLoading) {
|
|
70
|
+
return;
|
|
71
|
+
}
|
|
72
|
+
const newMessages = [...messages, message];
|
|
73
|
+
setMessages(newMessages);
|
|
74
|
+
return runChatCompletionAndHandleFunctionCall(newMessages);
|
|
75
|
+
});
|
|
76
|
+
const reload = () => __async(this, null, function* () {
|
|
77
|
+
if (isLoading || messages.length === 0) {
|
|
78
|
+
return;
|
|
79
|
+
}
|
|
80
|
+
let newMessages = [...messages];
|
|
81
|
+
const lastMessage = messages[messages.length - 1];
|
|
82
|
+
if (lastMessage.role === "assistant") {
|
|
83
|
+
newMessages = newMessages.slice(0, -1);
|
|
84
|
+
}
|
|
85
|
+
setMessages(newMessages);
|
|
86
|
+
return runChatCompletionAndHandleFunctionCall(newMessages);
|
|
87
|
+
});
|
|
88
|
+
const stop = () => {
|
|
89
|
+
throw new Error("Not implemented");
|
|
90
|
+
};
|
|
91
|
+
return {
|
|
92
|
+
messages,
|
|
93
|
+
append,
|
|
94
|
+
reload,
|
|
95
|
+
stop,
|
|
96
|
+
isLoading,
|
|
97
|
+
input,
|
|
98
|
+
setInput
|
|
99
|
+
};
|
|
100
|
+
}
|
|
101
|
+
|
|
102
|
+
export { useChat };
|
|
103
|
+
//# sourceMappingURL=out.js.map
|
|
104
|
+
//# sourceMappingURL=chunk-TDH7XE4X.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/hooks/use-chat.ts"],"names":["messages"],"mappings":";;;;;;;;;AAAA,SAAS,gBAAgB;AAEzB,SAAS,cAAc;AAgFhB,SAAS,QAAQ,SAA0D;AAChF,QAAM,CAAC,UAAU,WAAW,IAAI,SAAoB,CAAC,CAAC;AACtD,QAAM,CAAC,OAAO,QAAQ,IAAI,SAAS,EAAE;AACrC,QAAM,CAAC,WAAW,YAAY,IAAI,SAAS,KAAK;AAEhD,QAAM,oBAAoB,CAAOA,cAA0C;AACzE,WAAO,IAAI,QAAiB,CAAC,SAAS,WAAW;AAC/C,mBAAa,IAAI;AAEjB,YAAM,mBAA4B;AAAA,QAChC,IAAI,OAAO;AAAA,QACX,WAAW,oBAAI,KAAK;AAAA,QACpB,SAAS;AAAA,QACT,MAAM;AAAA,MACR;AAGA,kBAAY,CAAC,GAAGA,WAAU,mBAAK,iBAAkB,CAAC;AAElD,YAAM,sBAAsB,CAAC,GAAI,QAAQ,mBAAmB,CAAC,GAAI,GAAGA,SAAQ;AAE5E,YAAM,SAAS,IAAI,qBAAqB;AAAA,QACtC,KAAK,QAAQ;AAAA,MACf,CAAC;AAED,YAAM,UAAU,MAAM;AACpB,eAAO,IAAI,SAAS;AACpB,eAAO,IAAI,KAAK;AAChB,eAAO,IAAI,OAAO;AAClB,eAAO,IAAI,UAAU;AAAA,MACvB;AAEA,aAAO,GAAG,WAAW,CAAC,YAAY;AAChC,yBAAiB,WAAW;AAC5B,oBAAY,CAAC,GAAGA,WAAU,mBAAK,iBAAkB,CAAC;AAAA,MACpD,CAAC;AAED,aAAO,GAAG,OAAO,MAAM;AACrB,qBAAa,KAAK;AAClB,gBAAQ;AACR,gBAAQ,mBAAK,iBAAkB;AAAA,MACjC,CAAC;AAED,aAAO,GAAG,SAAS,CAAC,UAAU;AAC5B,qBAAa,KAAK;AAClB,gBAAQ;AACR,eAAO,KAAK;AAAA,MACd,CAAC;AAED,aAAO,GAAG,YAAY,CAAO,iBAAiB;AAC5C,yBAAiB,gBAAgB;AAAA,UAC/B,MAAM,aAAa;AAAA,UACnB,WAAW,KAAK,UAAU,aAAa,SAAS;AAAA,QAClD;AACA,oBAAY,CAAC,GAAGA,WAAU,mBAAK,iBAAkB,CAAC;AAElD,qBAAa,KAAK;AAClB,gBAAQ;AACR,gBAAQ,mBAAK,iBAAkB;AAAA,MACjC,EAAC;AAED,aAAO,MAAM;AAAA,QACX,eAAe,QAAQ;AAAA,QACvB,UAAU;AAAA,QACV,WAAW,QAAQ;AAAA,QACnB,SAAS,QAAQ;AAAA,MACnB,CAAC;AAAA,IACH,CAAC;AAAA,EACH;AAEA,QAAM,yCAAyC,CAAOA,cAAuC;AAC3F,UAAM,UAAU,MAAM,kBAAkBA,SAAQ;AAChD,QAAI,QAAQ,iBAAiB,QAAQ,gBAAgB;AACnD,YAAM,QAAQ,eAAeA,WAAU,QAAQ,aAAa;AAAA,IAC9D;AAAA,EACF;AAEA,QAAM,SAAS,CAAO,YAAoC;AACxD,QAAI,WAAW;AACb;AAAA,IACF;AACA,UAAM,cAAc,CAAC,GAAG,UAAU,OAAO;AACzC,gBAAY,WAAW;AACvB,WAAO,uCAAuC,WAAW;AAAA,EAC3D;AAEA,QAAM,SAAS,MAA2B;AACxC,QAAI,aAAa,SAAS,WAAW,GAAG;AACtC;AAAA,IACF;AACA,QAAI,cAAc,CAAC,GAAG,QAAQ;AAC9B,UAAM,cAAc,SAAS,SAAS,SAAS,CAAC;AAEhD,QAAI,YAAY,SAAS,aAAa;AACpC,oBAAc,YAAY,MAAM,GAAG,EAAE;AAAA,IACvC;AACA,gBAAY,WAAW;AAEvB,WAAO,uCAAuC,WAAW;AAAA,EAC3D;AAEA,QAAM,OAAO,MAAY;AACvB,UAAM,IAAI,MAAM,iBAAiB;AAAA,EACnC;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF","sourcesContent":["import { useState } from \"react\";\nimport { Message, Function, FunctionCallHandler } from \"../types\";\nimport { nanoid } from \"nanoid\";\nimport { ChatCompletionClient } from \"../openai/chat-completion-client\";\nimport { CopilotApiConfig } from \"../context\";\n\nexport type UseChatOptions = {\n /**\n * The API endpoint that accepts a `{ messages: Message[] }` object and returns\n * a stream of tokens of the AI chat response. Defaults to `/api/chat`.\n */\n api?: string;\n /**\n * A unique identifier for the chat. If not provided, a random one will be\n * generated. When provided, the `useChat` hook with the same `id` will\n * have shared states across components.\n */\n id?: string;\n /**\n * System messages of the chat. Defaults to an empty array.\n */\n initialMessages?: Message[];\n /**\n * Callback function to be called when a function call is received.\n * If the function returns a `ChatRequest` object, the request will be sent\n * automatically to the API and will be used to update the chat.\n */\n onFunctionCall?: FunctionCallHandler;\n /**\n * HTTP headers to be sent with the API request.\n */\n headers?: Record<string, string> | Headers;\n /**\n * Extra body object to be sent with the API request.\n * @example\n * Send a `sessionId` to the API along with the messages.\n * ```js\n * useChat({\n * body: {\n * sessionId: '123',\n * }\n * })\n * ```\n */\n body?: object;\n /**\n * Function definitions to be sent to the API.\n */\n functions?: Function[];\n};\n\nexport type UseChatHelpers = {\n /** Current messages in the chat */\n messages: Message[];\n /**\n * Append a user message to the chat list. This triggers the API call to fetch\n * the assistant's response.\n * @param message The message to append\n */\n append: (message: Message) => Promise<void>;\n /**\n * Reload the last AI chat response for the given chat history. If the last\n * message isn't from the assistant, it will request the API to generate a\n * new response.\n */\n reload: () => Promise<void>;\n /**\n * Abort the current request immediately, keep the generated tokens if any.\n */\n stop: () => void;\n /** The current value of the input */\n input: string;\n /** setState-powered method to update the input value */\n setInput: React.Dispatch<React.SetStateAction<string>>;\n /** Whether the API request is in progress */\n isLoading: boolean;\n};\n\nexport type UseChatOptionsWithCopilotConfig = UseChatOptions & {\n copilotConfig: CopilotApiConfig;\n};\n\nexport function useChat(options: UseChatOptionsWithCopilotConfig): UseChatHelpers {\n const [messages, setMessages] = useState<Message[]>([]);\n const [input, setInput] = useState(\"\");\n const [isLoading, setIsLoading] = useState(false);\n\n const runChatCompletion = async (messages: Message[]): Promise<Message> => {\n return new Promise<Message>((resolve, reject) => {\n setIsLoading(true);\n\n const assistantMessage: Message = {\n id: nanoid(),\n createdAt: new Date(),\n content: \"\",\n role: \"assistant\",\n };\n\n // Assistant messages are always copied when using setState\n setMessages([...messages, { ...assistantMessage }]);\n\n const messagesWithContext = [...(options.initialMessages || []), ...messages];\n\n const client = new ChatCompletionClient({\n url: options.api,\n });\n\n const cleanup = () => {\n client.off(\"content\");\n client.off(\"end\");\n client.off(\"error\");\n client.off(\"function\");\n };\n\n client.on(\"content\", (content) => {\n assistantMessage.content += content;\n setMessages([...messages, { ...assistantMessage }]);\n });\n\n client.on(\"end\", () => {\n setIsLoading(false);\n cleanup();\n resolve({ ...assistantMessage });\n });\n\n client.on(\"error\", (error) => {\n setIsLoading(false);\n cleanup();\n reject(error);\n });\n\n client.on(\"function\", async (functionCall) => {\n assistantMessage.function_call = {\n name: functionCall.name,\n arguments: JSON.stringify(functionCall.arguments),\n };\n setMessages([...messages, { ...assistantMessage }]);\n // quit early if we get a function call\n setIsLoading(false);\n cleanup();\n resolve({ ...assistantMessage });\n });\n\n client.fetch({\n copilotConfig: options.copilotConfig,\n messages: messagesWithContext,\n functions: options.functions,\n headers: options.headers,\n });\n });\n };\n\n const runChatCompletionAndHandleFunctionCall = async (messages: Message[]): Promise<void> => {\n const message = await runChatCompletion(messages);\n if (message.function_call && options.onFunctionCall) {\n await options.onFunctionCall(messages, message.function_call);\n }\n };\n\n const append = async (message: Message): Promise<void> => {\n if (isLoading) {\n return;\n }\n const newMessages = [...messages, message];\n setMessages(newMessages);\n return runChatCompletionAndHandleFunctionCall(newMessages);\n };\n\n const reload = async (): Promise<void> => {\n if (isLoading || messages.length === 0) {\n return;\n }\n let newMessages = [...messages];\n const lastMessage = messages[messages.length - 1];\n\n if (lastMessage.role === \"assistant\") {\n newMessages = newMessages.slice(0, -1);\n }\n setMessages(newMessages);\n\n return runChatCompletionAndHandleFunctionCall(newMessages);\n };\n\n const stop = (): void => {\n throw new Error(\"Not implemented\");\n };\n\n return {\n messages,\n append,\n reload,\n stop,\n isLoading,\n input,\n setInput,\n };\n}\n"]}
|
|
@@ -0,0 +1,177 @@
|
|
|
1
|
+
import { ChatCompletionTransport } from './chunk-VK25FIML.mjs';
|
|
2
|
+
import { __async, __spreadValues } from './chunk-MRXNTQOX.mjs';
|
|
3
|
+
import EventEmitter from 'eventemitter3';
|
|
4
|
+
|
|
5
|
+
var DEFAULT_MAX_TOKENS = 8192;
|
|
6
|
+
var ChatCompletionClient = class extends EventEmitter {
|
|
7
|
+
constructor(params) {
|
|
8
|
+
super();
|
|
9
|
+
this.chatCompletionTransport = null;
|
|
10
|
+
this.mode = null;
|
|
11
|
+
this.functionCallName = "";
|
|
12
|
+
this.functionCallArguments = "";
|
|
13
|
+
this.onData = (data) => {
|
|
14
|
+
if (this.mode === "function" && !data.choices[0].delta.function_call) {
|
|
15
|
+
const success = this.tryFlushFunctionCall();
|
|
16
|
+
if (!success) {
|
|
17
|
+
return;
|
|
18
|
+
}
|
|
19
|
+
}
|
|
20
|
+
this.mode = data.choices[0].delta.function_call ? "function" : "message";
|
|
21
|
+
if (this.mode === "message") {
|
|
22
|
+
if (data.choices[0].delta.content) {
|
|
23
|
+
this.emit("content", data.choices[0].delta.content);
|
|
24
|
+
}
|
|
25
|
+
return;
|
|
26
|
+
} else if (this.mode === "function") {
|
|
27
|
+
if (data.choices[0].delta.function_call.name) {
|
|
28
|
+
this.functionCallName = data.choices[0].delta.function_call.name;
|
|
29
|
+
}
|
|
30
|
+
if (data.choices[0].delta.function_call.arguments) {
|
|
31
|
+
this.functionCallArguments += data.choices[0].delta.function_call.arguments;
|
|
32
|
+
}
|
|
33
|
+
this.emit("partial", this.functionCallName, this.functionCallArguments);
|
|
34
|
+
return;
|
|
35
|
+
}
|
|
36
|
+
};
|
|
37
|
+
this.onError = (error) => {
|
|
38
|
+
this.emit("error", error);
|
|
39
|
+
this.cleanup();
|
|
40
|
+
};
|
|
41
|
+
this.onEnd = () => {
|
|
42
|
+
if (this.mode === "function") {
|
|
43
|
+
const success = this.tryFlushFunctionCall();
|
|
44
|
+
if (!success) {
|
|
45
|
+
return;
|
|
46
|
+
}
|
|
47
|
+
}
|
|
48
|
+
this.emit("end");
|
|
49
|
+
this.cleanup();
|
|
50
|
+
};
|
|
51
|
+
}
|
|
52
|
+
fetch(params) {
|
|
53
|
+
return __async(this, null, function* () {
|
|
54
|
+
params = __spreadValues({}, params);
|
|
55
|
+
if (params.model && params.model in maxTokensByModel) {
|
|
56
|
+
params.maxTokens || (params.maxTokens = maxTokensByModel[params.model]);
|
|
57
|
+
} else {
|
|
58
|
+
params.maxTokens || (params.maxTokens = DEFAULT_MAX_TOKENS);
|
|
59
|
+
}
|
|
60
|
+
params.functions || (params.functions = []);
|
|
61
|
+
params.messages = this.buildPrompt(params);
|
|
62
|
+
return yield this.runPrompt(params);
|
|
63
|
+
});
|
|
64
|
+
}
|
|
65
|
+
buildPrompt(params) {
|
|
66
|
+
let maxTokens = params.maxTokens;
|
|
67
|
+
const messages = params.messages;
|
|
68
|
+
const functions = params.functions;
|
|
69
|
+
const functionsNumTokens = countFunctionsTokens(functions);
|
|
70
|
+
if (functionsNumTokens > maxTokens) {
|
|
71
|
+
throw new Error(`Too many tokens in function calls: ${functionsNumTokens} > ${maxTokens}`);
|
|
72
|
+
}
|
|
73
|
+
maxTokens -= functionsNumTokens;
|
|
74
|
+
for (const message of messages) {
|
|
75
|
+
if (message.role === "system") {
|
|
76
|
+
const numTokens = this.countTokens(message);
|
|
77
|
+
maxTokens -= numTokens;
|
|
78
|
+
if (maxTokens < 0) {
|
|
79
|
+
throw new Error("Not enough tokens for system message.");
|
|
80
|
+
}
|
|
81
|
+
}
|
|
82
|
+
}
|
|
83
|
+
const result = [];
|
|
84
|
+
let cutoff = false;
|
|
85
|
+
const reversedMessages = [...messages].reverse();
|
|
86
|
+
for (const message of reversedMessages) {
|
|
87
|
+
if (message.role === "system") {
|
|
88
|
+
result.unshift(message);
|
|
89
|
+
continue;
|
|
90
|
+
} else if (cutoff) {
|
|
91
|
+
continue;
|
|
92
|
+
}
|
|
93
|
+
let numTokens = this.countTokens(message);
|
|
94
|
+
if (maxTokens < numTokens) {
|
|
95
|
+
cutoff = true;
|
|
96
|
+
continue;
|
|
97
|
+
}
|
|
98
|
+
result.unshift(message);
|
|
99
|
+
maxTokens -= numTokens;
|
|
100
|
+
}
|
|
101
|
+
return result;
|
|
102
|
+
}
|
|
103
|
+
runPrompt(params) {
|
|
104
|
+
return __async(this, null, function* () {
|
|
105
|
+
this.chatCompletionTransport = new ChatCompletionTransport({});
|
|
106
|
+
this.chatCompletionTransport.on("data", this.onData);
|
|
107
|
+
this.chatCompletionTransport.on("error", this.onError);
|
|
108
|
+
this.chatCompletionTransport.on("end", this.onEnd);
|
|
109
|
+
yield this.chatCompletionTransport.fetch(params);
|
|
110
|
+
});
|
|
111
|
+
}
|
|
112
|
+
tryFlushFunctionCall() {
|
|
113
|
+
let args = null;
|
|
114
|
+
try {
|
|
115
|
+
args = JSON.parse(this.functionCallArguments);
|
|
116
|
+
} catch (error) {
|
|
117
|
+
this.emit("error", error);
|
|
118
|
+
this.cleanup();
|
|
119
|
+
return false;
|
|
120
|
+
}
|
|
121
|
+
this.emit("function", {
|
|
122
|
+
name: this.functionCallName,
|
|
123
|
+
arguments: args
|
|
124
|
+
});
|
|
125
|
+
this.mode = null;
|
|
126
|
+
this.functionCallName = "";
|
|
127
|
+
this.functionCallArguments = "";
|
|
128
|
+
return true;
|
|
129
|
+
}
|
|
130
|
+
cleanup() {
|
|
131
|
+
if (this.chatCompletionTransport) {
|
|
132
|
+
this.chatCompletionTransport.off("data", this.onData);
|
|
133
|
+
this.chatCompletionTransport.off("error", this.onError);
|
|
134
|
+
this.chatCompletionTransport.off("end", this.onEnd);
|
|
135
|
+
}
|
|
136
|
+
this.chatCompletionTransport = null;
|
|
137
|
+
this.mode = null;
|
|
138
|
+
this.functionCallName = "";
|
|
139
|
+
this.functionCallArguments = "";
|
|
140
|
+
}
|
|
141
|
+
countTokens(message) {
|
|
142
|
+
if (message.content) {
|
|
143
|
+
return estimateTokens(message.content);
|
|
144
|
+
} else if (message.function_call) {
|
|
145
|
+
return estimateTokens(JSON.stringify(message.function_call));
|
|
146
|
+
}
|
|
147
|
+
return 0;
|
|
148
|
+
}
|
|
149
|
+
};
|
|
150
|
+
var maxTokensByModel = {
|
|
151
|
+
"gpt-3.5-turbo": 4097,
|
|
152
|
+
"gpt-3.5-turbo-16k": 16385,
|
|
153
|
+
"gpt-4": 8192,
|
|
154
|
+
"gpt-4-1106-preview": 8192,
|
|
155
|
+
"gpt-4-32k": 32768,
|
|
156
|
+
"gpt-3.5-turbo-0301": 4097,
|
|
157
|
+
"gpt-4-0314": 8192,
|
|
158
|
+
"gpt-4-32k-0314": 32768,
|
|
159
|
+
"gpt-3.5-turbo-0613": 4097,
|
|
160
|
+
"gpt-4-0613": 8192,
|
|
161
|
+
"gpt-4-32k-0613": 32768,
|
|
162
|
+
"gpt-3.5-turbo-16k-0613": 16385
|
|
163
|
+
};
|
|
164
|
+
function estimateTokens(text) {
|
|
165
|
+
return text.length / 3;
|
|
166
|
+
}
|
|
167
|
+
function countFunctionsTokens(functions) {
|
|
168
|
+
if (functions.length === 0) {
|
|
169
|
+
return 0;
|
|
170
|
+
}
|
|
171
|
+
const json = JSON.stringify(functions);
|
|
172
|
+
return estimateTokens(json);
|
|
173
|
+
}
|
|
174
|
+
|
|
175
|
+
export { ChatCompletionClient };
|
|
176
|
+
//# sourceMappingURL=out.js.map
|
|
177
|
+
//# sourceMappingURL=chunk-TKXPMMND.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/openai/chat-completion-client.ts"],"names":[],"mappings":";;;;;;;;;AAAA,OAAO,kBAAkB;AAiCzB,IAAM,qBAAqB;AAEpB,IAAM,uBAAN,cAAmC,aAAyC;AAAA,EAMjF,YAAY,QAA2C;AACrD,UAAM;AANR,SAAQ,0BAA0D;AAClE,SAAQ,OAAsC;AAC9C,SAAQ,mBAA2B;AACnC,SAAQ,wBAAgC;AAyExC,SAAQ,SAAS,CAAC,SAA8B;AAE9C,UAAI,KAAK,SAAS,cAAc,CAAC,KAAK,QAAQ,CAAC,EAAE,MAAM,eAAe;AACpE,cAAM,UAAU,KAAK,qBAAqB;AAC1C,YAAI,CAAC,SAAS;AACZ;AAAA,QACF;AAAA,MACF;AAEA,WAAK,OAAO,KAAK,QAAQ,CAAC,EAAE,MAAM,gBAAgB,aAAa;AAE/D,UAAI,KAAK,SAAS,WAAW;AAG3B,YAAI,KAAK,QAAQ,CAAC,EAAE,MAAM,SAAS;AACjC,eAAK,KAAK,WAAW,KAAK,QAAQ,CAAC,EAAE,MAAM,OAAO;AAAA,QACpD;AAEA;AAAA,MACF,WAAW,KAAK,SAAS,YAAY;AAGnC,YAAI,KAAK,QAAQ,CAAC,EAAE,MAAM,cAAe,MAAM;AAC7C,eAAK,mBAAmB,KAAK,QAAQ,CAAC,EAAE,MAAM,cAAe;AAAA,QAC/D;AACA,YAAI,KAAK,QAAQ,CAAC,EAAE,MAAM,cAAe,WAAW;AAClD,eAAK,yBAAyB,KAAK,QAAQ,CAAC,EAAE,MAAM,cAAe;AAAA,QACrE;AACA,aAAK,KAAK,WAAW,KAAK,kBAAkB,KAAK,qBAAqB;AAEtE;AAAA,MACF;AAAA,IACF;AAEA,SAAQ,UAAU,CAAC,UAAe;AAChC,WAAK,KAAK,SAAS,KAAK;AACxB,WAAK,QAAQ;AAAA,IACf;AAEA,SAAQ,QAAQ,MAAM;AACpB,UAAI,KAAK,SAAS,YAAY;AAC5B,cAAM,UAAU,KAAK,qBAAqB;AAC1C,YAAI,CAAC,SAAS;AACZ;AAAA,QACF;AAAA,MACF;AACA,WAAK,KAAK,KAAK;AACf,WAAK,QAAQ;AAAA,IACf;AAAA,EArHA;AAAA,EAEa,MAAM,QAA4C;AAAA;AAC7D,eAAS,mBAAK;AACd,UAAI,OAAO,SAAS,OAAO,SAAS,kBAAkB;AACpD,eAAO,cAAP,OAAO,YAAc,iBAAiB,OAAO,KAAK;AAAA,MACpD,OAAO;AACL,eAAO,cAAP,OAAO,YAAc;AAAA,MACvB;AAEA,aAAO,cAAP,OAAO,YAAc,CAAC;AACtB,aAAO,WAAW,KAAK,YAAY,MAAM;AACzC,aAAO,MAAM,KAAK,UAAU,MAAM;AAAA,IACpC;AAAA;AAAA,EAEQ,YAAY,QAAuD;AACzE,QAAI,YAAY,OAAO;AACvB,UAAM,WAAW,OAAO;AACxB,UAAM,YAAY,OAAO;AACzB,UAAM,qBAAqB,qBAAqB,SAAS;AACzD,QAAI,qBAAqB,WAAW;AAClC,YAAM,IAAI,MAAM,sCAAsC,wBAAwB,WAAW;AAAA,IAC3F;AACA,iBAAa;AAEb,eAAW,WAAW,UAAU;AAC9B,UAAI,QAAQ,SAAS,UAAU;AAC7B,cAAM,YAAY,KAAK,YAAY,OAAO;AAC1C,qBAAa;AAEb,YAAI,YAAY,GAAG;AACjB,gBAAM,IAAI,MAAM,uCAAuC;AAAA,QACzD;AAAA,MACF;AAAA,IACF;AAEA,UAAM,SAAoB,CAAC;AAC3B,QAAI,SAAkB;AAEtB,UAAM,mBAAmB,CAAC,GAAG,QAAQ,EAAE,QAAQ;AAC/C,eAAW,WAAW,kBAAkB;AACtC,UAAI,QAAQ,SAAS,UAAU;AAC7B,eAAO,QAAQ,OAAO;AACtB;AAAA,MACF,WAAW,QAAQ;AACjB;AAAA,MACF;AACA,UAAI,YAAY,KAAK,YAAY,OAAO;AACxC,UAAI,YAAY,WAAW;AACzB,iBAAS;AACT;AAAA,MACF;AACA,aAAO,QAAQ,OAAO;AACtB,mBAAa;AAAA,IACf;AAEA,WAAO;AAAA,EACT;AAAA,EAEc,UAAU,QAA2D;AAAA;AACjF,WAAK,0BAA0B,IAAI,wBAAwB,CAAC,CAAC;AAE7D,WAAK,wBAAwB,GAAG,QAAQ,KAAK,MAAM;AACnD,WAAK,wBAAwB,GAAG,SAAS,KAAK,OAAO;AACrD,WAAK,wBAAwB,GAAG,OAAO,KAAK,KAAK;AAEjD,YAAM,KAAK,wBAAwB,MAAM,MAAM;AAAA,IACjD;AAAA;AAAA,EAoDQ,uBAAgC;AACtC,QAAI,OAAY;AAChB,QAAI;AACF,aAAO,KAAK,MAAM,KAAK,qBAAqB;AAAA,IAC9C,SAAS,OAAP;AACA,WAAK,KAAK,SAAS,KAAK;AACxB,WAAK,QAAQ;AACb,aAAO;AAAA,IACT;AACA,SAAK,KAAK,YAAY;AAAA,MACpB,MAAM,KAAK;AAAA,MACX,WAAW;AAAA,IACb,CAAC;AACD,SAAK,OAAO;AACZ,SAAK,mBAAmB;AACxB,SAAK,wBAAwB;AAC7B,WAAO;AAAA,EACT;AAAA,EAEQ,UAAU;AAChB,QAAI,KAAK,yBAAyB;AAChC,WAAK,wBAAwB,IAAI,QAAQ,KAAK,MAAM;AACpD,WAAK,wBAAwB,IAAI,SAAS,KAAK,OAAO;AACtD,WAAK,wBAAwB,IAAI,OAAO,KAAK,KAAK;AAAA,IACpD;AACA,SAAK,0BAA0B;AAC/B,SAAK,OAAO;AACZ,SAAK,mBAAmB;AACxB,SAAK,wBAAwB;AAAA,EAC/B;AAAA,EAEO,YAAY,SAA0B;AAC3C,QAAI,QAAQ,SAAS;AACnB,aAAO,eAAe,QAAQ,OAAO;AAAA,IACvC,WAAW,QAAQ,eAAe;AAChC,aAAO,eAAe,KAAK,UAAU,QAAQ,aAAa,CAAC;AAAA,IAC7D;AACA,WAAO;AAAA,EACT;AACF;AAEA,IAAM,mBAA8C;AAAA,EAClD,iBAAiB;AAAA,EACjB,qBAAqB;AAAA,EACrB,SAAS;AAAA,EACT,sBAAsB;AAAA,EACtB,aAAa;AAAA,EACb,sBAAsB;AAAA,EACtB,cAAc;AAAA,EACd,kBAAkB;AAAA,EAClB,sBAAsB;AAAA,EACtB,cAAc;AAAA,EACd,kBAAkB;AAAA,EAClB,0BAA0B;AAC5B;AAEA,SAAS,eAAe,MAAsB;AAC5C,SAAO,KAAK,SAAS;AACvB;AAEA,SAAS,qBAAqB,WAA+B;AAC3D,MAAI,UAAU,WAAW,GAAG;AAC1B,WAAO;AAAA,EACT;AACA,QAAM,OAAO,KAAK,UAAU,SAAS;AACrC,SAAO,eAAe,IAAI;AAC5B","sourcesContent":["import EventEmitter from \"eventemitter3\";\nimport { Function, Message, Role } from \"../types\";\nimport {\n ChatCompletionTransport,\n ChatCompletionTransportFetchParams,\n} from \"./chat-completion-transport\";\n\ninterface ChatCompletionClientConfiguration {}\n\ninterface ChatCompletionClientEvents {\n content: string;\n partial: [string, string];\n error: any;\n function: {\n name: string;\n arguments: any;\n };\n end: void;\n}\n\nexport interface ChatCompletionChunk {\n choices: {\n delta: {\n role: Role;\n content?: string | null;\n function_call?: {\n name?: string;\n arguments?: string;\n };\n };\n }[];\n}\n\nconst DEFAULT_MAX_TOKENS = 8192;\n\nexport class ChatCompletionClient extends EventEmitter<ChatCompletionClientEvents> {\n private chatCompletionTransport: ChatCompletionTransport | null = null;\n private mode: \"function\" | \"message\" | null = null;\n private functionCallName: string = \"\";\n private functionCallArguments: string = \"\";\n\n constructor(params: ChatCompletionClientConfiguration) {\n super();\n }\n\n public async fetch(params: ChatCompletionTransportFetchParams) {\n params = { ...params };\n if (params.model && params.model in maxTokensByModel) {\n params.maxTokens ||= maxTokensByModel[params.model];\n } else {\n params.maxTokens ||= DEFAULT_MAX_TOKENS;\n }\n\n params.functions ||= [];\n params.messages = this.buildPrompt(params);\n return await this.runPrompt(params);\n }\n\n private buildPrompt(params: ChatCompletionTransportFetchParams): Message[] {\n let maxTokens = params.maxTokens!;\n const messages = params.messages!;\n const functions = params.functions!;\n const functionsNumTokens = countFunctionsTokens(functions);\n if (functionsNumTokens > maxTokens) {\n throw new Error(`Too many tokens in function calls: ${functionsNumTokens} > ${maxTokens}`);\n }\n maxTokens -= functionsNumTokens;\n\n for (const message of messages) {\n if (message.role === \"system\") {\n const numTokens = this.countTokens(message);\n maxTokens -= numTokens;\n\n if (maxTokens < 0) {\n throw new Error(\"Not enough tokens for system message.\");\n }\n }\n }\n\n const result: Message[] = [];\n let cutoff: boolean = false;\n\n const reversedMessages = [...messages].reverse();\n for (const message of reversedMessages) {\n if (message.role === \"system\") {\n result.unshift(message);\n continue;\n } else if (cutoff) {\n continue;\n }\n let numTokens = this.countTokens(message);\n if (maxTokens < numTokens) {\n cutoff = true;\n continue;\n }\n result.unshift(message);\n maxTokens -= numTokens;\n }\n\n return result;\n }\n\n private async runPrompt(params: ChatCompletionTransportFetchParams): Promise<void> {\n this.chatCompletionTransport = new ChatCompletionTransport({});\n\n this.chatCompletionTransport.on(\"data\", this.onData);\n this.chatCompletionTransport.on(\"error\", this.onError);\n this.chatCompletionTransport.on(\"end\", this.onEnd);\n\n await this.chatCompletionTransport.fetch(params);\n }\n\n private onData = (data: ChatCompletionChunk) => {\n // In case we are in a function call but the next message is not a function call, flush it.\n if (this.mode === \"function\" && !data.choices[0].delta.function_call) {\n const success = this.tryFlushFunctionCall();\n if (!success) {\n return;\n }\n }\n\n this.mode = data.choices[0].delta.function_call ? \"function\" : \"message\";\n\n if (this.mode === \"message\") {\n // if we get a message, emit the content and return;\n\n if (data.choices[0].delta.content) {\n this.emit(\"content\", data.choices[0].delta.content);\n }\n\n return;\n } else if (this.mode === \"function\") {\n // if we get a function call, we buffer the name and arguments, then emit a partial event.\n\n if (data.choices[0].delta.function_call!.name) {\n this.functionCallName = data.choices[0].delta.function_call!.name!;\n }\n if (data.choices[0].delta.function_call!.arguments) {\n this.functionCallArguments += data.choices[0].delta.function_call!.arguments!;\n }\n this.emit(\"partial\", this.functionCallName, this.functionCallArguments);\n\n return;\n }\n };\n\n private onError = (error: any) => {\n this.emit(\"error\", error);\n this.cleanup();\n };\n\n private onEnd = () => {\n if (this.mode === \"function\") {\n const success = this.tryFlushFunctionCall();\n if (!success) {\n return;\n }\n }\n this.emit(\"end\");\n this.cleanup();\n };\n\n private tryFlushFunctionCall(): boolean {\n let args: any = null;\n try {\n args = JSON.parse(this.functionCallArguments);\n } catch (error) {\n this.emit(\"error\", error);\n this.cleanup();\n return false;\n }\n this.emit(\"function\", {\n name: this.functionCallName,\n arguments: args,\n });\n this.mode = null;\n this.functionCallName = \"\";\n this.functionCallArguments = \"\";\n return true;\n }\n\n private cleanup() {\n if (this.chatCompletionTransport) {\n this.chatCompletionTransport.off(\"data\", this.onData);\n this.chatCompletionTransport.off(\"error\", this.onError);\n this.chatCompletionTransport.off(\"end\", this.onEnd);\n }\n this.chatCompletionTransport = null;\n this.mode = null;\n this.functionCallName = \"\";\n this.functionCallArguments = \"\";\n }\n\n public countTokens(message: Message): number {\n if (message.content) {\n return estimateTokens(message.content);\n } else if (message.function_call) {\n return estimateTokens(JSON.stringify(message.function_call));\n }\n return 0;\n }\n}\n\nconst maxTokensByModel: { [key: string]: number } = {\n \"gpt-3.5-turbo\": 4097,\n \"gpt-3.5-turbo-16k\": 16385,\n \"gpt-4\": 8192,\n \"gpt-4-1106-preview\": 8192,\n \"gpt-4-32k\": 32768,\n \"gpt-3.5-turbo-0301\": 4097,\n \"gpt-4-0314\": 8192,\n \"gpt-4-32k-0314\": 32768,\n \"gpt-3.5-turbo-0613\": 4097,\n \"gpt-4-0613\": 8192,\n \"gpt-4-32k-0613\": 32768,\n \"gpt-3.5-turbo-16k-0613\": 16385,\n};\n\nfunction estimateTokens(text: string): number {\n return text.length / 3;\n}\n\nfunction countFunctionsTokens(functions: Function[]): number {\n if (functions.length === 0) {\n return 0;\n }\n const json = JSON.stringify(functions);\n return estimateTokens(json);\n}\n"]}
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { processMessageStream } from './chunk-MZ5UN3BY.mjs';
|
|
2
|
-
import { defaultCopilotContextCategories } from './chunk-
|
|
3
|
-
import { CopilotContext
|
|
2
|
+
import { defaultCopilotContextCategories } from './chunk-QTET3FBV.mjs';
|
|
3
|
+
import { CopilotContext } from './chunk-6A4PCNMZ.mjs';
|
|
4
4
|
import { __async, __spreadValues } from './chunk-MRXNTQOX.mjs';
|
|
5
5
|
import { useContext, useState, useMemo } from 'react';
|
|
6
6
|
import { parseStreamPart } from '@copilotkit/shared';
|
|
@@ -38,7 +38,7 @@ function useCopilotChatV2(options) {
|
|
|
38
38
|
setStatus("in_progress");
|
|
39
39
|
setMessages((messages2) => [...messages2, { id: "", role: "user", content: input }]);
|
|
40
40
|
setInput("");
|
|
41
|
-
const apiUrl =
|
|
41
|
+
const apiUrl = copilotApiConfig.chatApiEndpointV2;
|
|
42
42
|
const functions = getChatCompletionFunctionDescriptions();
|
|
43
43
|
const result = yield fetch(apiUrl, {
|
|
44
44
|
method: "POST",
|
|
@@ -122,4 +122,4 @@ If you would like to call a function, call it without saying anything else.
|
|
|
122
122
|
|
|
123
123
|
export { defaultSystemMessage, useCopilotChatV2 };
|
|
124
124
|
//# sourceMappingURL=out.js.map
|
|
125
|
-
//# sourceMappingURL=chunk-
|
|
125
|
+
//# sourceMappingURL=chunk-TU4DNDO7.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/openai-assistants/hooks/use-copilot-chat-v2.ts"],"names":["messages","error"],"mappings":";;;;;;;;;;;;;;;AAAA,SAAS,YAAY,SAAS,gBAAgB;AAE9C,SAAkB,uBAAuB;AA6ClC,SAAS,iBAAiB,SAA0D;AACzF,QAAM;AAAA,IACJ;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF,IAAI,WAAW,cAAc;AAE7B,QAAM,CAAC,UAAU,WAAW,IAAI,SAAoB,CAAC,CAAC;AACtD,QAAM,CAAC,OAAO,QAAQ,IAAI,SAAS,EAAE;AACrC,QAAM,CAAC,UAAU,WAAW,IAAI,SAA6B,MAAS;AACtE,QAAM,CAAC,QAAQ,SAAS,IAAI,SAA0B,kBAAkB;AACxE,QAAM,CAAC,OAAO,QAAQ,IAAI,SAA8B,MAAS;AAEjE,QAAM,gBAAyB,QAAQ,MAAM;AAC3C,UAAM,qBAAqB,QAAQ,qBAAqB;AACxD,UAAM,gBAAgB,iBAAiB,CAAC,GAAG,+BAA+B;AAE1E,WAAO;AAAA,MACL,IAAI;AAAA,MACJ,SAAS,mBAAmB,aAAa;AAAA,MACzC,MAAM;AAAA,IACR;AAAA,EACF,GAAG,CAAC,kBAAkB,QAAQ,iBAAiB,CAAC;AAEhD,QAAM,oBAAoB,CAAC,MAAW;AACpC,aAAS,EAAE,OAAO,KAAK;AAAA,EACzB;AAEA,QAAM,gBAAgB,CAAO,MAAW;AA5E1C;AA6EI,MAAE,eAAe;AAEjB,QAAI,UAAU,IAAI;AAChB;AAAA,IACF;AAEA,cAAU,aAAa;AAEvB,gBAAY,CAACA,cAAa,CAAC,GAAGA,WAAU,EAAE,IAAI,IAAI,MAAM,QAAQ,SAAS,MAAM,CAAC,CAAC;AAEjF,aAAS,EAAE;AAEX,UAAM,SAAS,iBAAiB;AAEhC,UAAM,YAAY,sCAAsC;AAExD,UAAM,SAAS,MAAM,MAAM,QAAQ;AAAA,MACjC,QAAQ;AAAA,MACR,SAAS;AAAA,QACP,gBAAgB;AAAA,SACb,iBAAiB,UACjB,QAAQ;AAAA,MAEb,MAAM,KAAK,UAAU;AAAA;AAAA,QAEnB,WAAU,mBAAQ,aAAR,YAAoB,aAApB,YAAgC;AAAA,QAC1C,SAAS;AAAA,SACL,UAAU,SAAS,KAAK,EAAE,UAAqB,IAChD,iBAAiB,OACjB,QAAQ,KACZ;AAAA,IACH,CAAC;AAED,QAAI,OAAO,QAAQ,MAAM;AACvB,YAAM,IAAI,MAAM,6BAA6B;AAAA,IAC/C;AAEA,UAAM,qBAAqB,OAAO,KAAK,UAAU,GAAG,CAAC,YAAoB;AACvE,UAAI;AACF,cAAM,EAAE,MAAM,MAAM,IAAI,gBAAgB,OAAO;AAE/C,gBAAQ,MAAM;AAAA,UACZ,KAAK,qBAAqB;AAExB,wBAAY,CAACA,cAAa;AAAA,cACxB,GAAGA;AAAA,cACH;AAAA,gBACE,IAAI,MAAM;AAAA,gBACV,MAAM,MAAM;AAAA,gBACZ,SAAS,MAAM,QAAQ,CAAC,EAAE,KAAK;AAAA,cACjC;AAAA,YACF,CAAC;AACD;AAAA,UACF;AAAA,UAEA,KAAK,0BAA0B;AAC7B,wBAAY,MAAM,QAAQ;AAG1B,wBAAY,CAACA,cAAa;AACxB,oBAAM,cAAcA,UAASA,UAAS,SAAS,CAAC;AAChD,0BAAY,KAAK,MAAM;AACvB,qBAAO,CAAC,GAAGA,UAAS,MAAM,GAAGA,UAAS,SAAS,CAAC,GAAG,WAAW;AAAA,YAChE,CAAC;AAED;AAAA,UACF;AAAA,UAEA,KAAK,SAAS;AACZ,qBAAS,KAAK;AACd;AAAA,UACF;AAAA,QACF;AAAA,MACF,SAASC,QAAP;AACA,iBAASA,MAAK;AAAA,MAChB;AAAA,IACF,CAAC;AAED,cAAU,kBAAkB;AAAA,EAC9B;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;AAEO,SAAS,qBAAqB,eAA+B;AAClE,SAAO;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAQP;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAWF","sourcesContent":["import { useContext, useMemo, useState } from \"react\";\nimport { processMessageStream } from \"../utils\";\nimport { Message, parseStreamPart } from \"@copilotkit/shared\";\nimport { CopilotContext } from \"../../context\";\nimport { defaultCopilotContextCategories } from \"../../components\";\n\nexport type AssistantStatus = \"in_progress\" | \"awaiting_message\";\n\nexport interface RequestForwardingOptions {\n /**\n * The credentials mode to be used for the fetch request.\n * Possible values are: 'omit', 'same-origin', 'include'.\n * Defaults to 'same-origin'.\n */\n credentials?: RequestCredentials;\n /**\n * HTTP headers to be sent with the API request.\n */\n headers?: Record<string, string> | Headers;\n /**\n * Extra body object to be sent with the API request.\n * @example\n * Send a `sessionId` to the API along with the messages.\n * ```js\n * useChat({\n * body: {\n * sessionId: '123',\n * }\n * })\n * ```\n */\n body?: object;\n}\nexport interface UseCopilotChatOptionsV2 extends RequestForwardingOptions {\n makeSystemMessage?: (contextString: string) => string;\n threadId?: string | undefined;\n}\n\nexport interface UseCopilotChatV2Result {\n messages: Message[];\n input: string;\n handleInputChange: (e: any) => void;\n submitMessage: (e: any) => Promise<void>;\n status: AssistantStatus;\n error: unknown;\n}\n\nexport function useCopilotChatV2(options: UseCopilotChatOptionsV2): UseCopilotChatV2Result {\n const {\n getContextString,\n getChatCompletionFunctionDescriptions,\n getFunctionCallHandler,\n copilotApiConfig,\n } = useContext(CopilotContext);\n\n const [messages, setMessages] = useState<Message[]>([]);\n const [input, setInput] = useState(\"\");\n const [threadId, setThreadId] = useState<string | undefined>(undefined);\n const [status, setStatus] = useState<AssistantStatus>(\"awaiting_message\");\n const [error, setError] = useState<unknown | undefined>(undefined);\n\n const systemMessage: Message = useMemo(() => {\n const systemMessageMaker = options.makeSystemMessage || defaultSystemMessage;\n const contextString = getContextString([], defaultCopilotContextCategories); // TODO: make the context categories configurable\n\n return {\n id: \"system\",\n content: systemMessageMaker(contextString),\n role: \"system\",\n };\n }, [getContextString, options.makeSystemMessage]);\n\n const handleInputChange = (e: any) => {\n setInput(e.target.value);\n };\n\n const submitMessage = async (e: any) => {\n e.preventDefault();\n\n if (input === \"\") {\n return;\n }\n\n setStatus(\"in_progress\");\n\n setMessages((messages) => [...messages, { id: \"\", role: \"user\", content: input }]);\n\n setInput(\"\");\n\n const apiUrl = copilotApiConfig.chatApiEndpointV2;\n\n const functions = getChatCompletionFunctionDescriptions();\n\n const result = await fetch(apiUrl, {\n method: \"POST\",\n headers: {\n \"Content-Type\": \"application/json\",\n ...copilotApiConfig.headers,\n ...options.headers,\n },\n body: JSON.stringify({\n // always use user-provided threadId when available:\n threadId: options.threadId ?? threadId ?? null,\n message: input,\n ...(functions.length > 0 && { functions: functions }),\n ...copilotApiConfig.body,\n ...options.body,\n }),\n });\n\n if (result.body == null) {\n throw new Error(\"The response body is empty.\");\n }\n\n await processMessageStream(result.body.getReader(), (message: string) => {\n try {\n const { type, value } = parseStreamPart(message);\n\n switch (type) {\n case \"assistant_message\": {\n // append message:\n setMessages((messages) => [\n ...messages,\n {\n id: value.id,\n role: value.role,\n content: value.content[0].text.value,\n },\n ]);\n break;\n }\n\n case \"assistant_control_data\": {\n setThreadId(value.threadId);\n\n // set id of last message:\n setMessages((messages) => {\n const lastMessage = messages[messages.length - 1];\n lastMessage.id = value.messageId;\n return [...messages.slice(0, messages.length - 1), lastMessage];\n });\n\n break;\n }\n\n case \"error\": {\n setError(value);\n break;\n }\n }\n } catch (error) {\n setError(error);\n }\n });\n\n setStatus(\"awaiting_message\");\n };\n\n return {\n messages,\n input,\n handleInputChange,\n submitMessage,\n status,\n error,\n };\n}\n\nexport function defaultSystemMessage(contextString: string): string {\n return `\nPlease act as an efficient, competent, conscientious, and industrious professional assistant.\n\nHelp the user achieve their goals, and you do so in a way that is as efficient as possible, without unnecessary fluff, but also without sacrificing professionalism.\nAlways be polite and respectful, and prefer brevity over verbosity.\n\nThe user has provided you with the following context:\n\\`\\`\\`\n${contextString}\n\\`\\`\\`\n\nThey have also provided you with functions you can call to initiate actions on their behalf, or functions you can call to receive more information.\n\nPlease assist them as best you can.\n\nYou can ask them for clarifying questions if needed, but don't be annoying about it. If you can reasonably 'fill in the blanks' yourself, do so.\n\nIf you would like to call a function, call it without saying anything else.\n`;\n}\n"]}
|
|
@@ -0,0 +1,144 @@
|
|
|
1
|
+
import { __async, __spreadValues } from './chunk-MRXNTQOX.mjs';
|
|
2
|
+
import EventEmitter from 'eventemitter3';
|
|
3
|
+
|
|
4
|
+
var DEFAULT_MODEL = "gpt-4-1106-preview";
|
|
5
|
+
var ChatCompletionTransport = class extends EventEmitter {
|
|
6
|
+
constructor(params) {
|
|
7
|
+
super();
|
|
8
|
+
this.buffer = new Uint8Array();
|
|
9
|
+
this.bodyReader = null;
|
|
10
|
+
}
|
|
11
|
+
cleanup() {
|
|
12
|
+
return __async(this, null, function* () {
|
|
13
|
+
if (this.bodyReader) {
|
|
14
|
+
try {
|
|
15
|
+
yield this.bodyReader.cancel();
|
|
16
|
+
} catch (error) {
|
|
17
|
+
console.warn("Failed to cancel body reader:", error);
|
|
18
|
+
}
|
|
19
|
+
}
|
|
20
|
+
this.bodyReader = null;
|
|
21
|
+
this.buffer = new Uint8Array();
|
|
22
|
+
});
|
|
23
|
+
}
|
|
24
|
+
fetch(_0) {
|
|
25
|
+
return __async(this, arguments, function* ({
|
|
26
|
+
copilotConfig,
|
|
27
|
+
model,
|
|
28
|
+
messages,
|
|
29
|
+
functions,
|
|
30
|
+
temperature,
|
|
31
|
+
headers,
|
|
32
|
+
body,
|
|
33
|
+
signal
|
|
34
|
+
}) {
|
|
35
|
+
yield this.cleanup();
|
|
36
|
+
temperature || (temperature = 0.5);
|
|
37
|
+
functions || (functions = []);
|
|
38
|
+
model || (model = DEFAULT_MODEL);
|
|
39
|
+
const cleanedMessages = messages.map((message) => {
|
|
40
|
+
const { content, role, name, function_call } = message;
|
|
41
|
+
return { content, role, name, function_call };
|
|
42
|
+
});
|
|
43
|
+
try {
|
|
44
|
+
const response = yield fetch(copilotConfig.chatApiEndpoint, {
|
|
45
|
+
method: "POST",
|
|
46
|
+
headers: __spreadValues(__spreadValues({
|
|
47
|
+
"Content-Type": "application/json"
|
|
48
|
+
}, copilotConfig.headers), headers ? __spreadValues({}, headers) : {}),
|
|
49
|
+
body: JSON.stringify(__spreadValues(__spreadValues(__spreadValues(__spreadValues(__spreadValues({
|
|
50
|
+
model,
|
|
51
|
+
messages: cleanedMessages,
|
|
52
|
+
stream: true
|
|
53
|
+
}, functions.length ? { functions } : {}), temperature ? { temperature } : {}), functions.length != 0 ? { function_call: "auto" } : {}), copilotConfig.body), body ? __spreadValues({}, body) : {})),
|
|
54
|
+
signal
|
|
55
|
+
});
|
|
56
|
+
if (!response.ok) {
|
|
57
|
+
try {
|
|
58
|
+
const errorText = yield response.text();
|
|
59
|
+
yield this.cleanup();
|
|
60
|
+
const msg = `Error ${response.status}: ${errorText}`;
|
|
61
|
+
this.emit("error", new Error(msg));
|
|
62
|
+
} catch (_error) {
|
|
63
|
+
yield this.cleanup();
|
|
64
|
+
const msg = `Error ${response.status}: ${response.statusText}`;
|
|
65
|
+
this.emit("error", new Error(msg));
|
|
66
|
+
}
|
|
67
|
+
return;
|
|
68
|
+
}
|
|
69
|
+
if (response.body == null) {
|
|
70
|
+
yield this.cleanup();
|
|
71
|
+
const msg = "Response body is null";
|
|
72
|
+
this.emit("error", new Error(msg));
|
|
73
|
+
return;
|
|
74
|
+
}
|
|
75
|
+
this.bodyReader = response.body.getReader();
|
|
76
|
+
yield this.streamBody();
|
|
77
|
+
} catch (error) {
|
|
78
|
+
yield this.cleanup();
|
|
79
|
+
this.emit("error", error);
|
|
80
|
+
return;
|
|
81
|
+
}
|
|
82
|
+
});
|
|
83
|
+
}
|
|
84
|
+
streamBody() {
|
|
85
|
+
return __async(this, null, function* () {
|
|
86
|
+
while (true) {
|
|
87
|
+
try {
|
|
88
|
+
const { done, value } = yield this.bodyReader.read();
|
|
89
|
+
if (done) {
|
|
90
|
+
yield this.cleanup();
|
|
91
|
+
this.emit("end");
|
|
92
|
+
return;
|
|
93
|
+
}
|
|
94
|
+
const shouldContinue = yield this.processData(value);
|
|
95
|
+
if (!shouldContinue) {
|
|
96
|
+
return;
|
|
97
|
+
}
|
|
98
|
+
} catch (error) {
|
|
99
|
+
yield this.cleanup();
|
|
100
|
+
this.emit("error", error);
|
|
101
|
+
return;
|
|
102
|
+
}
|
|
103
|
+
}
|
|
104
|
+
});
|
|
105
|
+
}
|
|
106
|
+
processData(data) {
|
|
107
|
+
return __async(this, null, function* () {
|
|
108
|
+
const newBuffer = new Uint8Array(this.buffer.length + data.length);
|
|
109
|
+
newBuffer.set(this.buffer);
|
|
110
|
+
newBuffer.set(data, this.buffer.length);
|
|
111
|
+
this.buffer = newBuffer;
|
|
112
|
+
const dataString = new TextDecoder("utf-8").decode(this.buffer);
|
|
113
|
+
let lines = dataString.split("\n").filter((line) => line.trim() !== "");
|
|
114
|
+
if (!dataString.endsWith("\n")) {
|
|
115
|
+
const lastLine = lines.pop() || "";
|
|
116
|
+
const remainingBytes = new TextEncoder().encode(lastLine);
|
|
117
|
+
this.buffer = new Uint8Array(remainingBytes);
|
|
118
|
+
} else {
|
|
119
|
+
this.buffer = new Uint8Array();
|
|
120
|
+
}
|
|
121
|
+
for (const line of lines) {
|
|
122
|
+
const cleanedLine = line.replace(/^data: /, "");
|
|
123
|
+
if (cleanedLine === "[DONE]") {
|
|
124
|
+
yield this.cleanup();
|
|
125
|
+
this.emit("end");
|
|
126
|
+
return false;
|
|
127
|
+
}
|
|
128
|
+
let json;
|
|
129
|
+
try {
|
|
130
|
+
json = JSON.parse(cleanedLine);
|
|
131
|
+
} catch (error) {
|
|
132
|
+
console.error("Failed to parse JSON:", error);
|
|
133
|
+
continue;
|
|
134
|
+
}
|
|
135
|
+
this.emit("data", json);
|
|
136
|
+
}
|
|
137
|
+
return true;
|
|
138
|
+
});
|
|
139
|
+
}
|
|
140
|
+
};
|
|
141
|
+
|
|
142
|
+
export { ChatCompletionTransport };
|
|
143
|
+
//# sourceMappingURL=out.js.map
|
|
144
|
+
//# sourceMappingURL=chunk-VK25FIML.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../src/openai/chat-completion-transport.ts"],"names":[],"mappings":";;;;;;AAAA,OAAO,kBAAkB;AAwBzB,IAAM,gBAAgB;AAEf,IAAM,0BAAN,cAAsC,aAA4C;AAAA,EAIvF,YAAY,QAA8C;AACxD,UAAM;AAJR,SAAQ,SAAS,IAAI,WAAW;AAChC,SAAQ,aAA6D;AAAA,EAIrE;AAAA,EAEc,UAAU;AAAA;AACtB,UAAI,KAAK,YAAY;AACnB,YAAI;AACF,gBAAM,KAAK,WAAW,OAAO;AAAA,QAC/B,SAAS,OAAP;AACA,kBAAQ,KAAK,iCAAiC,KAAK;AAAA,QACrD;AAAA,MACF;AACA,WAAK,aAAa;AAClB,WAAK,SAAS,IAAI,WAAW;AAAA,IAC/B;AAAA;AAAA,EAEa,MAAM,IASmC;AAAA,+CATnC;AAAA,MACjB;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF,GAAsD;AACpD,YAAM,KAAK,QAAQ;AAEnB,oCAAgB;AAChB,gCAAc,CAAC;AACf,wBAAU;AAGV,YAAM,kBAAkB,SAAS,IAAI,CAAC,YAAY;AAChD,cAAM,EAAE,SAAS,MAAM,MAAM,cAAc,IAAI;AAC/C,eAAO,EAAE,SAAS,MAAM,MAAM,cAAc;AAAA,MAC9C,CAAC;AAED,UAAI;AACF,cAAM,WAAW,MAAM,MAAM,cAAc,iBAAiB;AAAA,UAC1D,QAAQ;AAAA,UACR,SAAS;AAAA,YACP,gBAAgB;AAAA,aACb,cAAc,UACb,UAAU,mBAAK,WAAY,CAAC;AAAA,UAElC,MAAM,KAAK,UAAU;AAAA,YACnB;AAAA,YACA,UAAU;AAAA,YACV,QAAQ;AAAA,aACJ,UAAU,SAAS,EAAE,UAAU,IAAI,CAAC,IACpC,cAAc,EAAE,YAAY,IAAI,CAAC,IACjC,UAAU,UAAU,IAAI,EAAE,eAAe,OAAO,IAAI,CAAC,IACtD,cAAc,OACb,OAAO,mBAAK,QAAS,CAAC,EAC3B;AAAA,UACD;AAAA,QACF,CAAC;AAED,YAAI,CAAC,SAAS,IAAI;AAChB,cAAI;AACF,kBAAM,YAAY,MAAM,SAAS,KAAK;AACtC,kBAAM,KAAK,QAAQ;AACnB,kBAAM,MAAM,SAAS,SAAS,WAAW;AACzC,iBAAK,KAAK,SAAS,IAAI,MAAM,GAAG,CAAC;AAAA,UACnC,SAAS,QAAP;AACA,kBAAM,KAAK,QAAQ;AACnB,kBAAM,MAAM,SAAS,SAAS,WAAW,SAAS;AAClD,iBAAK,KAAK,SAAS,IAAI,MAAM,GAAG,CAAC;AAAA,UACnC;AACA;AAAA,QACF;AAEA,YAAI,SAAS,QAAQ,MAAM;AACzB,gBAAM,KAAK,QAAQ;AACnB,gBAAM,MAAM;AACZ,eAAK,KAAK,SAAS,IAAI,MAAM,GAAG,CAAC;AACjC;AAAA,QACF;AAEA,aAAK,aAAa,SAAS,KAAK,UAAU;AAE1C,cAAM,KAAK,WAAW;AAAA,MACxB,SAAS,OAAP;AACA,cAAM,KAAK,QAAQ;AACnB,aAAK,KAAK,SAAS,KAAK;AACxB;AAAA,MACF;AAAA,IACF;AAAA;AAAA,EAEc,aAAa;AAAA;AACzB,aAAO,MAAM;AACX,YAAI;AACF,gBAAM,EAAE,MAAM,MAAM,IAAI,MAAM,KAAK,WAAY,KAAK;AAEpD,cAAI,MAAM;AACR,kBAAM,KAAK,QAAQ;AACnB,iBAAK,KAAK,KAAK;AACf;AAAA,UACF;AAEA,gBAAM,iBAAiB,MAAM,KAAK,YAAY,KAAK;AAEnD,cAAI,CAAC,gBAAgB;AACnB;AAAA,UACF;AAAA,QACF,SAAS,OAAP;AACA,gBAAM,KAAK,QAAQ;AACnB,eAAK,KAAK,SAAS,KAAK;AACxB;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAAA;AAAA,EAEc,YAAY,MAAoC;AAAA;AAE5D,YAAM,YAAY,IAAI,WAAW,KAAK,OAAO,SAAS,KAAK,MAAM;AACjE,gBAAU,IAAI,KAAK,MAAM;AACzB,gBAAU,IAAI,MAAM,KAAK,OAAO,MAAM;AACtC,WAAK,SAAS;AAEd,YAAM,aAAa,IAAI,YAAY,OAAO,EAAE,OAAO,KAAK,MAAM;AAE9D,UAAI,QAAQ,WAAW,MAAM,IAAI,EAAE,OAAO,CAAC,SAAS,KAAK,KAAK,MAAM,EAAE;AAGtE,UAAI,CAAC,WAAW,SAAS,IAAI,GAAG;AAC9B,cAAM,WAAW,MAAM,IAAI,KAAK;AAChC,cAAM,iBAAiB,IAAI,YAAY,EAAE,OAAO,QAAQ;AACxD,aAAK,SAAS,IAAI,WAAW,cAAc;AAAA,MAC7C,OAAO;AACL,aAAK,SAAS,IAAI,WAAW;AAAA,MAC/B;AAEA,iBAAW,QAAQ,OAAO;AACxB,cAAM,cAAc,KAAK,QAAQ,WAAW,EAAE;AAE9C,YAAI,gBAAgB,UAAU;AAC5B,gBAAM,KAAK,QAAQ;AACnB,eAAK,KAAK,KAAK;AACf,iBAAO;AAAA,QACT;AAEA,YAAI;AACJ,YAAI;AACF,iBAAO,KAAK,MAAM,WAAW;AAAA,QAC/B,SAAS,OAAP;AACA,kBAAQ,MAAM,yBAAyB,KAAK;AAC5C;AAAA,QACF;AAEA,aAAK,KAAK,QAAQ,IAAI;AAAA,MACxB;AACA,aAAO;AAAA,IACT;AAAA;AACF","sourcesContent":["import EventEmitter from \"eventemitter3\";\nimport { Message, Function } from \"../types\";\nimport { CopilotApiConfig } from \"../context\";\n\nexport interface ChatCompletionTransportConfiguration {}\n\ninterface ChatCompletionTransportEvents {\n end: void;\n data: any;\n error: any;\n}\n\nexport interface ChatCompletionTransportFetchParams {\n copilotConfig: CopilotApiConfig;\n model?: string;\n messages: Message[];\n functions?: Function[];\n temperature?: number;\n maxTokens?: number;\n headers?: Record<string, string> | Headers;\n body?: object;\n signal?: AbortSignal;\n}\n\nconst DEFAULT_MODEL = \"gpt-4-1106-preview\";\n\nexport class ChatCompletionTransport extends EventEmitter<ChatCompletionTransportEvents> {\n private buffer = new Uint8Array();\n private bodyReader: ReadableStreamDefaultReader<Uint8Array> | null = null;\n\n constructor(params: ChatCompletionTransportConfiguration) {\n super();\n }\n\n private async cleanup() {\n if (this.bodyReader) {\n try {\n await this.bodyReader.cancel();\n } catch (error) {\n console.warn(\"Failed to cancel body reader:\", error);\n }\n }\n this.bodyReader = null;\n this.buffer = new Uint8Array();\n }\n\n public async fetch({\n copilotConfig,\n model,\n messages,\n functions,\n temperature,\n headers,\n body,\n signal,\n }: ChatCompletionTransportFetchParams): Promise<void> {\n await this.cleanup();\n\n temperature ||= 0.5;\n functions ||= [];\n model ||= DEFAULT_MODEL;\n\n // clean up any extra properties from messages\n const cleanedMessages = messages.map((message) => {\n const { content, role, name, function_call } = message;\n return { content, role, name, function_call };\n });\n\n try {\n const response = await fetch(copilotConfig.chatApiEndpoint, {\n method: \"POST\",\n headers: {\n \"Content-Type\": \"application/json\",\n ...copilotConfig.headers,\n ...(headers ? { ...headers } : {}),\n },\n body: JSON.stringify({\n model,\n messages: cleanedMessages,\n stream: true,\n ...(functions.length ? { functions } : {}),\n ...(temperature ? { temperature } : {}),\n ...(functions.length != 0 ? { function_call: \"auto\" } : {}),\n ...copilotConfig.body,\n ...(body ? { ...body } : {}),\n }),\n signal,\n });\n\n if (!response.ok) {\n try {\n const errorText = await response.text();\n await this.cleanup();\n const msg = `Error ${response.status}: ${errorText}`;\n this.emit(\"error\", new Error(msg));\n } catch (_error) {\n await this.cleanup();\n const msg = `Error ${response.status}: ${response.statusText}`;\n this.emit(\"error\", new Error(msg));\n }\n return;\n }\n\n if (response.body == null) {\n await this.cleanup();\n const msg = \"Response body is null\";\n this.emit(\"error\", new Error(msg));\n return;\n }\n\n this.bodyReader = response.body.getReader();\n\n await this.streamBody();\n } catch (error) {\n await this.cleanup();\n this.emit(\"error\", error);\n return;\n }\n }\n\n private async streamBody() {\n while (true) {\n try {\n const { done, value } = await this.bodyReader!.read();\n\n if (done) {\n await this.cleanup();\n this.emit(\"end\");\n return;\n }\n\n const shouldContinue = await this.processData(value);\n\n if (!shouldContinue) {\n return;\n }\n } catch (error) {\n await this.cleanup();\n this.emit(\"error\", error);\n return;\n }\n }\n }\n\n private async processData(data: Uint8Array): Promise<boolean> {\n // Append new data to the temp buffer\n const newBuffer = new Uint8Array(this.buffer.length + data.length);\n newBuffer.set(this.buffer);\n newBuffer.set(data, this.buffer.length);\n this.buffer = newBuffer;\n\n const dataString = new TextDecoder(\"utf-8\").decode(this.buffer);\n\n let lines = dataString.split(\"\\n\").filter((line) => line.trim() !== \"\");\n\n // If the last line isn't complete, keep it in the buffer for next time\n if (!dataString.endsWith(\"\\n\")) {\n const lastLine = lines.pop() || \"\";\n const remainingBytes = new TextEncoder().encode(lastLine);\n this.buffer = new Uint8Array(remainingBytes);\n } else {\n this.buffer = new Uint8Array();\n }\n\n for (const line of lines) {\n const cleanedLine = line.replace(/^data: /, \"\");\n\n if (cleanedLine === \"[DONE]\") {\n await this.cleanup();\n this.emit(\"end\");\n return false;\n }\n\n let json;\n try {\n json = JSON.parse(cleanedLine);\n } catch (error) {\n console.error(\"Failed to parse JSON:\", error);\n continue;\n }\n\n this.emit(\"data\", json);\n }\n return true;\n }\n}\n"]}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":[],"names":[],"mappings":""}
|
|
@@ -1,10 +1,10 @@
|
|
|
1
1
|
import { ReactNode } from 'react';
|
|
2
2
|
import { CopilotApiConfig } from '../../context/copilot-context.js';
|
|
3
|
-
import 'ai';
|
|
4
|
-
import '../../hooks/use-tree.js';
|
|
5
3
|
import '../../types/annotated-function.js';
|
|
6
|
-
import 'openai/resources/chat';
|
|
7
4
|
import '../../types/document-pointer.js';
|
|
5
|
+
import '../../types/base.js';
|
|
6
|
+
import '../../hooks/use-tree.js';
|
|
7
|
+
import 'openai/resources/chat';
|
|
8
8
|
|
|
9
9
|
/**
|
|
10
10
|
* Props for the CopilotProvider when using a chat API endpoint.
|
|
@@ -1,11 +1,11 @@
|
|
|
1
1
|
import { CopilotProviderProps } from './copilot-provider-props.js';
|
|
2
2
|
import 'react';
|
|
3
3
|
import '../../context/copilot-context.js';
|
|
4
|
-
import 'ai';
|
|
5
|
-
import '../../hooks/use-tree.js';
|
|
6
4
|
import '../../types/annotated-function.js';
|
|
7
|
-
import 'openai/resources/chat';
|
|
8
5
|
import '../../types/document-pointer.js';
|
|
6
|
+
import '../../types/base.js';
|
|
7
|
+
import '../../hooks/use-tree.js';
|
|
8
|
+
import 'openai/resources/chat';
|
|
9
9
|
|
|
10
10
|
/**
|
|
11
11
|
* The CopilotProvider component.
|
|
@@ -1,8 +1,8 @@
|
|
|
1
|
-
export { CopilotProvider, defaultCopilotContextCategories } from '../../chunk-
|
|
1
|
+
export { CopilotProvider, defaultCopilotContextCategories } from '../../chunk-QTET3FBV.mjs';
|
|
2
2
|
import '../../chunk-F2JIAPZQ.mjs';
|
|
3
3
|
import '../../chunk-VUY2K2DI.mjs';
|
|
4
|
+
import '../../chunk-6A4PCNMZ.mjs';
|
|
4
5
|
import '../../chunk-YULKJPY3.mjs';
|
|
5
|
-
import '../../chunk-YGJFU4ZP.mjs';
|
|
6
6
|
import '../../chunk-MRXNTQOX.mjs';
|
|
7
7
|
//# sourceMappingURL=out.js.map
|
|
8
8
|
//# sourceMappingURL=copilot-provider.mjs.map
|
|
@@ -2,8 +2,8 @@ export { CopilotProvider, defaultCopilotContextCategories } from './copilot-prov
|
|
|
2
2
|
export { CopilotProviderApiConfigProps, CopilotProviderApiEndpointProps, CopilotProviderProps } from './copilot-provider-props.js';
|
|
3
3
|
import 'react';
|
|
4
4
|
import '../../context/copilot-context.js';
|
|
5
|
-
import 'ai';
|
|
6
|
-
import '../../hooks/use-tree.js';
|
|
7
5
|
import '../../types/annotated-function.js';
|
|
8
|
-
import 'openai/resources/chat';
|
|
9
6
|
import '../../types/document-pointer.js';
|
|
7
|
+
import '../../types/base.js';
|
|
8
|
+
import '../../hooks/use-tree.js';
|
|
9
|
+
import 'openai/resources/chat';
|