@copilotkit/react-core 0.16.0-alpha.3 → 0.17.0-alpha.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.turbo/turbo-build.log +159 -125
- package/CHANGELOG.md +4 -36
- package/dist/components/copilot-provider/copilot-provider-props.d.ts +3 -3
- package/dist/components/copilot-provider/copilot-provider-props.mjs +1 -2
- package/dist/components/copilot-provider/copilot-provider-props.mjs.map +1 -1
- package/dist/components/copilot-provider/copilot-provider.d.ts +11 -6
- package/dist/components/copilot-provider/copilot-provider.mjs +468 -7
- package/dist/components/copilot-provider/copilot-provider.mjs.map +1 -1
- package/dist/components/copilot-provider/index.d.ts +3 -3
- package/dist/components/copilot-provider/index.mjs +467 -8
- package/dist/components/copilot-provider/index.mjs.map +1 -1
- package/dist/components/copilot-provider/{standard-cpilot-api-config.d.ts → standard-copilot-api-config.d.ts} +3 -3
- package/dist/{chunk-6N6X7K7T.mjs → components/copilot-provider/standard-copilot-api-config.mjs} +5 -5
- package/dist/components/copilot-provider/standard-copilot-api-config.mjs.map +1 -0
- package/dist/components/index.d.ts +3 -3
- package/dist/components/index.mjs +467 -9
- package/dist/components/index.mjs.map +1 -1
- package/dist/context/copilot-context.d.ts +6 -10
- package/dist/context/copilot-context.mjs +70 -3
- package/dist/context/copilot-context.mjs.map +1 -1
- package/dist/context/index.d.ts +4 -4
- package/dist/context/index.mjs +69 -4
- package/dist/context/index.mjs.map +1 -1
- package/dist/hooks/index.d.ts +6 -1
- package/dist/hooks/index.mjs +654 -14
- package/dist/hooks/index.mjs.map +1 -1
- package/dist/hooks/use-chat.d.ts +84 -0
- package/dist/hooks/use-chat.mjs +461 -0
- package/dist/hooks/use-chat.mjs.map +1 -0
- package/dist/hooks/use-copilot-chat.d.ts +10 -3
- package/dist/hooks/use-copilot-chat.mjs +599 -10
- package/dist/hooks/use-copilot-chat.mjs.map +1 -1
- package/dist/hooks/use-flat-category-store.mjs +68 -3
- package/dist/hooks/use-flat-category-store.mjs.map +1 -1
- package/dist/hooks/use-make-copilot-actionable.mjs +95 -4
- package/dist/hooks/use-make-copilot-actionable.mjs.map +1 -1
- package/dist/hooks/use-make-copilot-document-readable.mjs +87 -4
- package/dist/hooks/use-make-copilot-document-readable.mjs.map +1 -1
- package/dist/hooks/use-make-copilot-readable.mjs +87 -4
- package/dist/hooks/use-make-copilot-readable.mjs.map +1 -1
- package/dist/hooks/use-tree.mjs +153 -3
- package/dist/hooks/use-tree.mjs.map +1 -1
- package/dist/index.d.ts +6 -2
- package/dist/index.mjs +1204 -21
- package/dist/index.mjs.map +1 -1
- package/dist/openai/chat-completion-client.d.ts +56 -0
- package/dist/openai/chat-completion-client.mjs +360 -0
- package/dist/openai/chat-completion-client.mjs.map +1 -0
- package/dist/openai/chat-completion-stream.d.ts +21 -0
- package/dist/openai/chat-completion-stream.mjs +221 -0
- package/dist/openai/chat-completion-stream.mjs.map +1 -0
- package/dist/openai/chat-completion-transport.d.ts +40 -0
- package/dist/openai/chat-completion-transport.mjs +181 -0
- package/dist/openai/chat-completion-transport.mjs.map +1 -0
- package/dist/openai/index.d.ts +10 -0
- package/dist/openai/index.mjs +221 -0
- package/dist/openai/index.mjs.map +1 -0
- package/dist/openai-assistants/hooks/index.mjs +235 -14
- package/dist/openai-assistants/hooks/index.mjs.map +1 -1
- package/dist/openai-assistants/hooks/use-assistants.mjs +52 -8
- package/dist/openai-assistants/hooks/use-assistants.mjs.map +1 -1
- package/dist/openai-assistants/hooks/use-copilot-chat-v2.mjs +236 -13
- package/dist/openai-assistants/hooks/use-copilot-chat-v2.mjs.map +1 -1
- package/dist/openai-assistants/index.mjs +236 -15
- package/dist/openai-assistants/index.mjs.map +1 -1
- package/dist/openai-assistants/utils/index.mjs +46 -4
- package/dist/openai-assistants/utils/index.mjs.map +1 -1
- package/dist/openai-assistants/utils/process-message-stream.mjs +46 -3
- package/dist/openai-assistants/utils/process-message-stream.mjs.map +1 -1
- package/dist/types/annotated-function.mjs +0 -2
- package/dist/types/annotated-function.mjs.map +1 -1
- package/dist/types/base.d.ts +56 -0
- package/dist/types/base.mjs +1 -0
- package/dist/types/base.mjs.map +1 -0
- package/dist/types/document-pointer.mjs +0 -2
- package/dist/types/document-pointer.mjs.map +1 -1
- package/dist/types/index.d.ts +1 -0
- package/dist/types/index.mjs +0 -2
- package/dist/types/index.mjs.map +1 -1
- package/dist/types/message.d.ts +2 -0
- package/dist/types/message.mjs +1 -0
- package/dist/types/message.mjs.map +1 -0
- package/dist/utils/utils.mjs +0 -2
- package/dist/utils/utils.mjs.map +1 -1
- package/dist/utils/utils.test.mjs +0 -1
- package/dist/utils/utils.test.mjs.map +1 -1
- package/package.json +3 -3
- package/src/components/copilot-provider/copilot-provider.tsx +14 -9
- package/src/context/copilot-context.tsx +3 -14
- package/src/context/index.ts +0 -1
- package/src/hooks/index.ts +1 -0
- package/src/hooks/use-chat.ts +197 -0
- package/src/hooks/use-copilot-chat.ts +10 -22
- package/src/index.tsx +1 -0
- package/src/openai/chat-completion-client.ts +240 -0
- package/src/openai/chat-completion-stream.ts +56 -0
- package/src/openai/chat-completion-transport.ts +190 -0
- package/src/openai/index.tsx +5 -0
- package/src/openai-assistants/hooks/use-copilot-chat-v2.ts +2 -2
- package/src/types/base.ts +61 -0
- package/src/types/index.ts +1 -0
- package/src/types/message.ts +0 -0
- package/dist/chunk-3MAIWZNZ.mjs +0 -58
- package/dist/chunk-3MAIWZNZ.mjs.map +0 -1
- package/dist/chunk-45PUEKTG.mjs +0 -19
- package/dist/chunk-45PUEKTG.mjs.map +0 -1
- package/dist/chunk-6LKBKYRJ.mjs +0 -165
- package/dist/chunk-6LKBKYRJ.mjs.map +0 -1
- package/dist/chunk-6LNDDH6K.mjs +0 -19
- package/dist/chunk-6LNDDH6K.mjs.map +0 -1
- package/dist/chunk-6N6X7K7T.mjs.map +0 -1
- package/dist/chunk-7GFKOIO7.mjs +0 -3
- package/dist/chunk-7GFKOIO7.mjs.map +0 -1
- package/dist/chunk-BABVSMJR.mjs +0 -3
- package/dist/chunk-BABVSMJR.mjs.map +0 -1
- package/dist/chunk-E3P5YZO2.mjs +0 -27
- package/dist/chunk-E3P5YZO2.mjs.map +0 -1
- package/dist/chunk-EFZPSZWO.mjs +0 -3
- package/dist/chunk-EFZPSZWO.mjs.map +0 -1
- package/dist/chunk-FRAKUJWH.mjs +0 -3
- package/dist/chunk-FRAKUJWH.mjs.map +0 -1
- package/dist/chunk-JD7BAH7U.mjs +0 -3
- package/dist/chunk-JD7BAH7U.mjs.map +0 -1
- package/dist/chunk-JHJ7LUTD.mjs +0 -125
- package/dist/chunk-JHJ7LUTD.mjs.map +0 -1
- package/dist/chunk-MRXNTQOX.mjs +0 -55
- package/dist/chunk-MRXNTQOX.mjs.map +0 -1
- package/dist/chunk-MZ5UN3BY.mjs +0 -28
- package/dist/chunk-MZ5UN3BY.mjs.map +0 -1
- package/dist/chunk-OFRZZ5OF.mjs +0 -79
- package/dist/chunk-OFRZZ5OF.mjs.map +0 -1
- package/dist/chunk-QACD2U6P.mjs +0 -3
- package/dist/chunk-QACD2U6P.mjs.map +0 -1
- package/dist/chunk-SPCZTZCY.mjs +0 -3
- package/dist/chunk-SPCZTZCY.mjs.map +0 -1
- package/dist/chunk-VUY2K2DI.mjs +0 -135
- package/dist/chunk-VUY2K2DI.mjs.map +0 -1
- package/dist/chunk-YPSGKPDA.mjs +0 -3
- package/dist/chunk-YPSGKPDA.mjs.map +0 -1
- package/dist/chunk-YULKJPY3.mjs +0 -70
- package/dist/chunk-YULKJPY3.mjs.map +0 -1
- package/dist/components/copilot-provider/standard-cpilot-api-config.mjs +0 -4
- package/dist/components/copilot-provider/standard-cpilot-api-config.mjs.map +0 -1
- /package/src/components/copilot-provider/{standard-cpilot-api-config.tsx → standard-copilot-api-config.tsx} +0 -0
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../../src/hooks/use-chat.ts","../../src/openai/chat-completion-client.ts","../../src/openai/chat-completion-transport.ts"],"sourcesContent":["import { useState } from \"react\";\nimport { Message, Function, FunctionCallHandler } from \"../types\";\nimport { nanoid } from \"nanoid\";\nimport { ChatCompletionClient } from \"../openai/chat-completion-client\";\nimport { CopilotApiConfig } from \"../context\";\n\nexport type UseChatOptions = {\n /**\n * The API endpoint that accepts a `{ messages: Message[] }` object and returns\n * a stream of tokens of the AI chat response. Defaults to `/api/chat`.\n */\n api?: string;\n /**\n * A unique identifier for the chat. If not provided, a random one will be\n * generated. When provided, the `useChat` hook with the same `id` will\n * have shared states across components.\n */\n id?: string;\n /**\n * System messages of the chat. Defaults to an empty array.\n */\n initialMessages?: Message[];\n /**\n * Callback function to be called when a function call is received.\n * If the function returns a `ChatRequest` object, the request will be sent\n * automatically to the API and will be used to update the chat.\n */\n onFunctionCall?: FunctionCallHandler;\n /**\n * HTTP headers to be sent with the API request.\n */\n headers?: Record<string, string> | Headers;\n /**\n * Extra body object to be sent with the API request.\n * @example\n * Send a `sessionId` to the API along with the messages.\n * ```js\n * useChat({\n * body: {\n * sessionId: '123',\n * }\n * })\n * ```\n */\n body?: object;\n /**\n * Function definitions to be sent to the API.\n */\n functions?: Function[];\n};\n\nexport type UseChatHelpers = {\n /** Current messages in the chat */\n messages: Message[];\n /**\n * Append a user message to the chat list. This triggers the API call to fetch\n * the assistant's response.\n * @param message The message to append\n */\n append: (message: Message) => Promise<void>;\n /**\n * Reload the last AI chat response for the given chat history. If the last\n * message isn't from the assistant, it will request the API to generate a\n * new response.\n */\n reload: () => Promise<void>;\n /**\n * Abort the current request immediately, keep the generated tokens if any.\n */\n stop: () => void;\n /** The current value of the input */\n input: string;\n /** setState-powered method to update the input value */\n setInput: React.Dispatch<React.SetStateAction<string>>;\n /** Whether the API request is in progress */\n isLoading: boolean;\n};\n\nexport type UseChatOptionsWithCopilotConfig = UseChatOptions & {\n copilotConfig: CopilotApiConfig;\n};\n\nexport function useChat(options: UseChatOptionsWithCopilotConfig): UseChatHelpers {\n const [messages, setMessages] = useState<Message[]>([]);\n const [input, setInput] = useState(\"\");\n const [isLoading, setIsLoading] = useState(false);\n\n const runChatCompletion = async (messages: Message[]): Promise<Message> => {\n return new Promise<Message>((resolve, reject) => {\n setIsLoading(true);\n\n const assistantMessage: Message = {\n id: nanoid(),\n createdAt: new Date(),\n content: \"\",\n role: \"assistant\",\n };\n\n // Assistant messages are always copied when using setState\n setMessages([...messages, { ...assistantMessage }]);\n\n const messagesWithContext = [...(options.initialMessages || []), ...messages];\n\n const client = new ChatCompletionClient({\n url: options.api || \"/api/copilotkit/openai\",\n });\n\n const cleanup = () => {\n client.off(\"content\");\n client.off(\"end\");\n client.off(\"error\");\n client.off(\"function\");\n };\n\n client.on(\"content\", (content) => {\n assistantMessage.content += content;\n setMessages([...messages, { ...assistantMessage }]);\n });\n\n client.on(\"end\", () => {\n setIsLoading(false);\n cleanup();\n resolve({ ...assistantMessage });\n });\n\n client.on(\"error\", (error) => {\n setIsLoading(false);\n cleanup();\n reject(error);\n });\n\n client.on(\"function\", async (functionCall) => {\n assistantMessage.function_call = {\n name: functionCall.name,\n arguments: JSON.stringify(functionCall.arguments),\n };\n setMessages([...messages, { ...assistantMessage }]);\n // quit early if we get a function call\n setIsLoading(false);\n cleanup();\n resolve({ ...assistantMessage });\n });\n\n client.fetch({\n messages: messagesWithContext,\n functions: options.functions,\n headers: options.headers,\n copilotConfig: options.copilotConfig,\n });\n });\n };\n\n const runChatCompletionAndHandleFunctionCall = async (messages: Message[]): Promise<void> => {\n const message = await runChatCompletion(messages);\n if (message.function_call && options.onFunctionCall) {\n await options.onFunctionCall(messages, message.function_call);\n }\n };\n\n const append = async (message: Message): Promise<void> => {\n if (isLoading) {\n return;\n }\n const newMessages = [...messages, message];\n setMessages(newMessages);\n return runChatCompletionAndHandleFunctionCall(newMessages);\n };\n\n const reload = async (): Promise<void> => {\n if (isLoading || messages.length === 0) {\n return;\n }\n let newMessages = [...messages];\n const lastMessage = messages[messages.length - 1];\n\n if (lastMessage.role === \"assistant\") {\n newMessages = newMessages.slice(0, -1);\n }\n setMessages(newMessages);\n\n return runChatCompletionAndHandleFunctionCall(newMessages);\n };\n\n const stop = (): void => {\n throw new Error(\"Not implemented\");\n };\n\n return {\n messages,\n append,\n reload,\n stop,\n isLoading,\n input,\n setInput,\n };\n}\n","import EventEmitter from \"eventemitter3\";\nimport { Function, Message, Role } from \"../types\";\nimport {\n ChatCompletionTransport,\n ChatCompletionTransportFetchParams,\n} from \"./chat-completion-transport\";\n\ninterface ChatCompletionClientConfiguration {\n url: string;\n model?: string;\n}\n\ninterface ChatCompletionClientEvents {\n content: string;\n partial: [string, string];\n error: any;\n function: {\n name: string;\n arguments: any;\n };\n end: void;\n}\n\nexport interface ChatCompletionChunk {\n choices: {\n delta: {\n role: Role;\n content?: string | null;\n function_call?: {\n name?: string;\n arguments?: string;\n };\n };\n }[];\n}\n\nconst DEFAULT_MAX_TOKENS = 8192;\n\nexport class ChatCompletionClient extends EventEmitter<ChatCompletionClientEvents> {\n private url: string;\n private model?: string;\n\n private chatCompletionTransport: ChatCompletionTransport | null = null;\n private mode: \"function\" | \"message\" | null = null;\n private functionCallName: string = \"\";\n private functionCallArguments: string = \"\";\n\n constructor(params: ChatCompletionClientConfiguration) {\n super();\n this.model = params.model;\n this.url = params.url;\n }\n\n public async fetch(params: ChatCompletionTransportFetchParams) {\n params = { ...params };\n if (this.model && this.model in maxTokensByModel) {\n params.maxTokens ||= maxTokensByModel[this.model];\n } else {\n params.maxTokens ||= DEFAULT_MAX_TOKENS;\n }\n\n params.functions ||= [];\n params.model = this.model;\n params.messages = this.buildPrompt(params);\n return await this.runPrompt(params);\n }\n\n private buildPrompt(params: ChatCompletionTransportFetchParams): Message[] {\n let maxTokens = params.maxTokens!;\n const messages = params.messages!;\n const functions = params.functions!;\n const functionsNumTokens = countFunctionsTokens(functions);\n if (functionsNumTokens > maxTokens) {\n throw new Error(`Too many tokens in function calls: ${functionsNumTokens} > ${maxTokens}`);\n }\n maxTokens -= functionsNumTokens;\n\n for (const message of messages) {\n if (message.role === \"system\") {\n const numTokens = this.countTokens(message);\n maxTokens -= numTokens;\n\n if (maxTokens < 0) {\n throw new Error(\"Not enough tokens for system message.\");\n }\n }\n }\n\n const result: Message[] = [];\n let cutoff: boolean = false;\n\n const reversedMessages = [...messages].reverse();\n for (const message of reversedMessages) {\n if (message.role === \"system\") {\n result.unshift(message);\n continue;\n } else if (cutoff) {\n continue;\n }\n let numTokens = this.countTokens(message);\n if (maxTokens < numTokens) {\n cutoff = true;\n continue;\n }\n result.unshift(message);\n maxTokens -= numTokens;\n }\n\n return result;\n }\n\n private async runPrompt(params: ChatCompletionTransportFetchParams): Promise<void> {\n this.chatCompletionTransport = new ChatCompletionTransport({\n url: this.url,\n });\n\n this.chatCompletionTransport.on(\"data\", this.onData);\n this.chatCompletionTransport.on(\"error\", this.onError);\n this.chatCompletionTransport.on(\"end\", this.onEnd);\n\n await this.chatCompletionTransport.fetch(params);\n }\n\n private onData = (data: ChatCompletionChunk) => {\n // In case we are in a function call but the next message is not a function call, flush it.\n if (this.mode === \"function\" && !data.choices[0].delta.function_call) {\n const success = this.tryFlushFunctionCall();\n if (!success) {\n return;\n }\n }\n\n this.mode = data.choices[0].delta.function_call ? \"function\" : \"message\";\n\n if (this.mode === \"message\") {\n // if we get a message, emit the content and return;\n\n if (data.choices[0].delta.content) {\n this.emit(\"content\", data.choices[0].delta.content);\n }\n\n return;\n } else if (this.mode === \"function\") {\n // if we get a function call, we buffer the name and arguments, then emit a partial event.\n\n if (data.choices[0].delta.function_call!.name) {\n this.functionCallName = data.choices[0].delta.function_call!.name!;\n }\n if (data.choices[0].delta.function_call!.arguments) {\n this.functionCallArguments += data.choices[0].delta.function_call!.arguments!;\n }\n this.emit(\"partial\", this.functionCallName, this.functionCallArguments);\n\n return;\n }\n };\n\n private onError = (error: any) => {\n this.emit(\"error\", error);\n this.cleanup();\n };\n\n private onEnd = () => {\n if (this.mode === \"function\") {\n const success = this.tryFlushFunctionCall();\n if (!success) {\n return;\n }\n }\n this.emit(\"end\");\n this.cleanup();\n };\n\n private tryFlushFunctionCall(): boolean {\n let args: any = null;\n try {\n args = JSON.parse(this.functionCallArguments);\n } catch (error) {\n this.emit(\"error\", error);\n this.cleanup();\n return false;\n }\n this.emit(\"function\", {\n name: this.functionCallName,\n arguments: args,\n });\n this.mode = null;\n this.functionCallName = \"\";\n this.functionCallArguments = \"\";\n return true;\n }\n\n private cleanup() {\n if (this.chatCompletionTransport) {\n this.chatCompletionTransport.off(\"data\", this.onData);\n this.chatCompletionTransport.off(\"error\", this.onError);\n this.chatCompletionTransport.off(\"end\", this.onEnd);\n }\n this.chatCompletionTransport = null;\n this.mode = null;\n this.functionCallName = \"\";\n this.functionCallArguments = \"\";\n }\n\n public countTokens(message: Message): number {\n if (message.content) {\n return estimateTokens(message.content);\n } else if (message.function_call) {\n return estimateTokens(JSON.stringify(message.function_call));\n }\n return 0;\n }\n}\n\nconst maxTokensByModel: { [key: string]: number } = {\n \"gpt-3.5-turbo\": 4097,\n \"gpt-3.5-turbo-16k\": 16385,\n \"gpt-4\": 8192,\n \"gpt-4-1106-preview\": 8192,\n \"gpt-4-32k\": 32768,\n \"gpt-3.5-turbo-0301\": 4097,\n \"gpt-4-0314\": 8192,\n \"gpt-4-32k-0314\": 32768,\n \"gpt-3.5-turbo-0613\": 4097,\n \"gpt-4-0613\": 8192,\n \"gpt-4-32k-0613\": 32768,\n \"gpt-3.5-turbo-16k-0613\": 16385,\n};\n\nfunction estimateTokens(text: string): number {\n return text.length / 3;\n}\n\nfunction countFunctionsTokens(functions: Function[]): number {\n if (functions.length === 0) {\n return 0;\n }\n const json = JSON.stringify(functions);\n return estimateTokens(json);\n}\n","import EventEmitter from \"eventemitter3\";\nimport { Message, Function } from \"../types\";\nimport { CopilotApiConfig } from \"../context\";\n\nexport interface ChatCompletionTransportConfiguration {\n url: string;\n}\n\ninterface ChatCompletionTransportEvents {\n end: void;\n data: any;\n error: any;\n}\n\nexport interface ChatCompletionTransportFetchParams {\n model?: string;\n messages: Message[];\n functions?: Function[];\n temperature?: number;\n maxTokens?: number;\n headers?: Record<string, string> | Headers;\n body?: object;\n copilotConfig: CopilotApiConfig;\n signal?: AbortSignal;\n}\n\nconst DEFAULT_MODEL = \"gpt-4-1106-preview\";\n\nexport class ChatCompletionTransport extends EventEmitter<ChatCompletionTransportEvents> {\n private buffer = new Uint8Array();\n private bodyReader: ReadableStreamDefaultReader<Uint8Array> | null = null;\n private url: string;\n\n constructor({ url }: ChatCompletionTransportConfiguration) {\n super();\n this.url = url;\n }\n\n private async cleanup() {\n if (this.bodyReader) {\n try {\n await this.bodyReader.cancel();\n } catch (error) {\n console.warn(\"Failed to cancel body reader:\", error);\n }\n }\n this.bodyReader = null;\n this.buffer = new Uint8Array();\n }\n\n public async fetch({\n model,\n messages,\n copilotConfig,\n functions,\n temperature,\n headers,\n body,\n signal,\n }: ChatCompletionTransportFetchParams): Promise<void> {\n await this.cleanup();\n\n temperature ||= 0.5;\n functions ||= [];\n model ||= DEFAULT_MODEL;\n\n // clean up any extra properties from messages\n const cleanedMessages = messages.map((message) => {\n const { content, role, name, function_call } = message;\n return { content, role, name, function_call };\n });\n\n try {\n const response = await fetch(this.url, {\n method: \"POST\",\n headers: {\n \"Content-Type\": \"application/json\",\n ...copilotConfig.headers,\n ...(headers ? { ...headers } : {}),\n },\n body: JSON.stringify({\n model,\n messages: cleanedMessages,\n stream: true,\n ...(functions.length ? { functions } : {}),\n ...(temperature ? { temperature } : {}),\n ...(functions.length != 0 ? { function_call: \"auto\" } : {}),\n ...copilotConfig.body,\n ...(body ? { ...body } : {}),\n }),\n signal,\n });\n\n if (!response.ok) {\n try {\n const errorText = await response.text();\n await this.cleanup();\n const msg = `Error ${response.status}: ${errorText}`;\n this.emit(\"error\", new Error(msg));\n } catch (_error) {\n await this.cleanup();\n const msg = `Error ${response.status}: ${response.statusText}`;\n this.emit(\"error\", new Error(msg));\n }\n return;\n }\n\n if (response.body == null) {\n await this.cleanup();\n const msg = \"Response body is null\";\n this.emit(\"error\", new Error(msg));\n return;\n }\n\n this.bodyReader = response.body.getReader();\n\n await this.streamBody();\n } catch (error) {\n await this.cleanup();\n this.emit(\"error\", error);\n return;\n }\n }\n\n private async streamBody() {\n while (true) {\n try {\n const { done, value } = await this.bodyReader!.read();\n\n if (done) {\n await this.cleanup();\n this.emit(\"end\");\n return;\n }\n\n const shouldContinue = await this.processData(value);\n\n if (!shouldContinue) {\n return;\n }\n } catch (error) {\n await this.cleanup();\n this.emit(\"error\", error);\n return;\n }\n }\n }\n\n private async processData(data: Uint8Array): Promise<boolean> {\n // Append new data to the temp buffer\n const newBuffer = new Uint8Array(this.buffer.length + data.length);\n newBuffer.set(this.buffer);\n newBuffer.set(data, this.buffer.length);\n this.buffer = newBuffer;\n\n const dataString = new TextDecoder(\"utf-8\").decode(this.buffer);\n\n let lines = dataString.split(\"\\n\").filter((line) => line.trim() !== \"\");\n\n // If the last line isn't complete, keep it in the buffer for next time\n if (!dataString.endsWith(\"\\n\")) {\n const lastLine = lines.pop() || \"\";\n const remainingBytes = new TextEncoder().encode(lastLine);\n this.buffer = new Uint8Array(remainingBytes);\n } else {\n this.buffer = new Uint8Array();\n }\n\n for (const line of lines) {\n const cleanedLine = line.replace(/^data: /, \"\");\n\n if (cleanedLine === \"[DONE]\") {\n await this.cleanup();\n this.emit(\"end\");\n return false;\n }\n\n let json;\n try {\n json = JSON.parse(cleanedLine);\n } catch (error) {\n console.error(\"Failed to parse JSON:\", error);\n continue;\n }\n\n this.emit(\"data\", json);\n }\n return true;\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA,SAAS,gBAAgB;AAEzB,SAAS,cAAc;;;ACFvB,OAAOA,mBAAkB;;;ACAzB,OAAO,kBAAkB;AA0BzB,IAAM,gBAAgB;AAEf,IAAM,0BAAN,cAAsC,aAA4C;AAAA,EAKvF,YAAY,EAAE,IAAI,GAAyC;AACzD,UAAM;AALR,SAAQ,SAAS,IAAI,WAAW;AAChC,SAAQ,aAA6D;AAKnE,SAAK,MAAM;AAAA,EACb;AAAA,EAEc,UAAU;AAAA;AACtB,UAAI,KAAK,YAAY;AACnB,YAAI;AACF,gBAAM,KAAK,WAAW,OAAO;AAAA,QAC/B,SAAS,OAAP;AACA,kBAAQ,KAAK,iCAAiC,KAAK;AAAA,QACrD;AAAA,MACF;AACA,WAAK,aAAa;AAClB,WAAK,SAAS,IAAI,WAAW;AAAA,IAC/B;AAAA;AAAA,EAEa,MAAM,IASmC;AAAA,+CATnC;AAAA,MACjB;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF,GAAsD;AACpD,YAAM,KAAK,QAAQ;AAEnB,oCAAgB;AAChB,gCAAc,CAAC;AACf,wBAAU;AAGV,YAAM,kBAAkB,SAAS,IAAI,CAAC,YAAY;AAChD,cAAM,EAAE,SAAS,MAAM,MAAM,cAAc,IAAI;AAC/C,eAAO,EAAE,SAAS,MAAM,MAAM,cAAc;AAAA,MAC9C,CAAC;AAED,UAAI;AACF,cAAM,WAAW,MAAM,MAAM,KAAK,KAAK;AAAA,UACrC,QAAQ;AAAA,UACR,SAAS;AAAA,YACP,gBAAgB;AAAA,aACb,cAAc,UACb,UAAU,mBAAK,WAAY,CAAC;AAAA,UAElC,MAAM,KAAK,UAAU;AAAA,YACnB;AAAA,YACA,UAAU;AAAA,YACV,QAAQ;AAAA,aACJ,UAAU,SAAS,EAAE,UAAU,IAAI,CAAC,IACpC,cAAc,EAAE,YAAY,IAAI,CAAC,IACjC,UAAU,UAAU,IAAI,EAAE,eAAe,OAAO,IAAI,CAAC,IACtD,cAAc,OACb,OAAO,mBAAK,QAAS,CAAC,EAC3B;AAAA,UACD;AAAA,QACF,CAAC;AAED,YAAI,CAAC,SAAS,IAAI;AAChB,cAAI;AACF,kBAAM,YAAY,MAAM,SAAS,KAAK;AACtC,kBAAM,KAAK,QAAQ;AACnB,kBAAM,MAAM,SAAS,SAAS,WAAW;AACzC,iBAAK,KAAK,SAAS,IAAI,MAAM,GAAG,CAAC;AAAA,UACnC,SAAS,QAAP;AACA,kBAAM,KAAK,QAAQ;AACnB,kBAAM,MAAM,SAAS,SAAS,WAAW,SAAS;AAClD,iBAAK,KAAK,SAAS,IAAI,MAAM,GAAG,CAAC;AAAA,UACnC;AACA;AAAA,QACF;AAEA,YAAI,SAAS,QAAQ,MAAM;AACzB,gBAAM,KAAK,QAAQ;AACnB,gBAAM,MAAM;AACZ,eAAK,KAAK,SAAS,IAAI,MAAM,GAAG,CAAC;AACjC;AAAA,QACF;AAEA,aAAK,aAAa,SAAS,KAAK,UAAU;AAE1C,cAAM,KAAK,WAAW;AAAA,MACxB,SAAS,OAAP;AACA,cAAM,KAAK,QAAQ;AACnB,aAAK,KAAK,SAAS,KAAK;AACxB;AAAA,MACF;AAAA,IACF;AAAA;AAAA,EAEc,aAAa;AAAA;AACzB,aAAO,MAAM;AACX,YAAI;AACF,gBAAM,EAAE,MAAM,MAAM,IAAI,MAAM,KAAK,WAAY,KAAK;AAEpD,cAAI,MAAM;AACR,kBAAM,KAAK,QAAQ;AACnB,iBAAK,KAAK,KAAK;AACf;AAAA,UACF;AAEA,gBAAM,iBAAiB,MAAM,KAAK,YAAY,KAAK;AAEnD,cAAI,CAAC,gBAAgB;AACnB;AAAA,UACF;AAAA,QACF,SAAS,OAAP;AACA,gBAAM,KAAK,QAAQ;AACnB,eAAK,KAAK,SAAS,KAAK;AACxB;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAAA;AAAA,EAEc,YAAY,MAAoC;AAAA;AAE5D,YAAM,YAAY,IAAI,WAAW,KAAK,OAAO,SAAS,KAAK,MAAM;AACjE,gBAAU,IAAI,KAAK,MAAM;AACzB,gBAAU,IAAI,MAAM,KAAK,OAAO,MAAM;AACtC,WAAK,SAAS;AAEd,YAAM,aAAa,IAAI,YAAY,OAAO,EAAE,OAAO,KAAK,MAAM;AAE9D,UAAI,QAAQ,WAAW,MAAM,IAAI,EAAE,OAAO,CAAC,SAAS,KAAK,KAAK,MAAM,EAAE;AAGtE,UAAI,CAAC,WAAW,SAAS,IAAI,GAAG;AAC9B,cAAM,WAAW,MAAM,IAAI,KAAK;AAChC,cAAM,iBAAiB,IAAI,YAAY,EAAE,OAAO,QAAQ;AACxD,aAAK,SAAS,IAAI,WAAW,cAAc;AAAA,MAC7C,OAAO;AACL,aAAK,SAAS,IAAI,WAAW;AAAA,MAC/B;AAEA,iBAAW,QAAQ,OAAO;AACxB,cAAM,cAAc,KAAK,QAAQ,WAAW,EAAE;AAE9C,YAAI,gBAAgB,UAAU;AAC5B,gBAAM,KAAK,QAAQ;AACnB,eAAK,KAAK,KAAK;AACf,iBAAO;AAAA,QACT;AAEA,YAAI;AACJ,YAAI;AACF,iBAAO,KAAK,MAAM,WAAW;AAAA,QAC/B,SAAS,OAAP;AACA,kBAAQ,MAAM,yBAAyB,KAAK;AAC5C;AAAA,QACF;AAEA,aAAK,KAAK,QAAQ,IAAI;AAAA,MACxB;AACA,aAAO;AAAA,IACT;AAAA;AACF;;;ADzJA,IAAM,qBAAqB;AAEpB,IAAM,uBAAN,cAAmCC,cAAyC;AAAA,EASjF,YAAY,QAA2C;AACrD,UAAM;AANR,SAAQ,0BAA0D;AAClE,SAAQ,OAAsC;AAC9C,SAAQ,mBAA2B;AACnC,SAAQ,wBAAgC;AA8ExC,SAAQ,SAAS,CAAC,SAA8B;AAE9C,UAAI,KAAK,SAAS,cAAc,CAAC,KAAK,QAAQ,CAAC,EAAE,MAAM,eAAe;AACpE,cAAM,UAAU,KAAK,qBAAqB;AAC1C,YAAI,CAAC,SAAS;AACZ;AAAA,QACF;AAAA,MACF;AAEA,WAAK,OAAO,KAAK,QAAQ,CAAC,EAAE,MAAM,gBAAgB,aAAa;AAE/D,UAAI,KAAK,SAAS,WAAW;AAG3B,YAAI,KAAK,QAAQ,CAAC,EAAE,MAAM,SAAS;AACjC,eAAK,KAAK,WAAW,KAAK,QAAQ,CAAC,EAAE,MAAM,OAAO;AAAA,QACpD;AAEA;AAAA,MACF,WAAW,KAAK,SAAS,YAAY;AAGnC,YAAI,KAAK,QAAQ,CAAC,EAAE,MAAM,cAAe,MAAM;AAC7C,eAAK,mBAAmB,KAAK,QAAQ,CAAC,EAAE,MAAM,cAAe;AAAA,QAC/D;AACA,YAAI,KAAK,QAAQ,CAAC,EAAE,MAAM,cAAe,WAAW;AAClD,eAAK,yBAAyB,KAAK,QAAQ,CAAC,EAAE,MAAM,cAAe;AAAA,QACrE;AACA,aAAK,KAAK,WAAW,KAAK,kBAAkB,KAAK,qBAAqB;AAEtE;AAAA,MACF;AAAA,IACF;AAEA,SAAQ,UAAU,CAAC,UAAe;AAChC,WAAK,KAAK,SAAS,KAAK;AACxB,WAAK,QAAQ;AAAA,IACf;AAEA,SAAQ,QAAQ,MAAM;AACpB,UAAI,KAAK,SAAS,YAAY;AAC5B,cAAM,UAAU,KAAK,qBAAqB;AAC1C,YAAI,CAAC,SAAS;AACZ;AAAA,QACF;AAAA,MACF;AACA,WAAK,KAAK,KAAK;AACf,WAAK,QAAQ;AAAA,IACf;AA1HE,SAAK,QAAQ,OAAO;AACpB,SAAK,MAAM,OAAO;AAAA,EACpB;AAAA,EAEa,MAAM,QAA4C;AAAA;AAC7D,eAAS,mBAAK;AACd,UAAI,KAAK,SAAS,KAAK,SAAS,kBAAkB;AAChD,eAAO,cAAP,OAAO,YAAc,iBAAiB,KAAK,KAAK;AAAA,MAClD,OAAO;AACL,eAAO,cAAP,OAAO,YAAc;AAAA,MACvB;AAEA,aAAO,cAAP,OAAO,YAAc,CAAC;AACtB,aAAO,QAAQ,KAAK;AACpB,aAAO,WAAW,KAAK,YAAY,MAAM;AACzC,aAAO,MAAM,KAAK,UAAU,MAAM;AAAA,IACpC;AAAA;AAAA,EAEQ,YAAY,QAAuD;AACzE,QAAI,YAAY,OAAO;AACvB,UAAM,WAAW,OAAO;AACxB,UAAM,YAAY,OAAO;AACzB,UAAM,qBAAqB,qBAAqB,SAAS;AACzD,QAAI,qBAAqB,WAAW;AAClC,YAAM,IAAI,MAAM,sCAAsC,wBAAwB,WAAW;AAAA,IAC3F;AACA,iBAAa;AAEb,eAAW,WAAW,UAAU;AAC9B,UAAI,QAAQ,SAAS,UAAU;AAC7B,cAAM,YAAY,KAAK,YAAY,OAAO;AAC1C,qBAAa;AAEb,YAAI,YAAY,GAAG;AACjB,gBAAM,IAAI,MAAM,uCAAuC;AAAA,QACzD;AAAA,MACF;AAAA,IACF;AAEA,UAAM,SAAoB,CAAC;AAC3B,QAAI,SAAkB;AAEtB,UAAM,mBAAmB,CAAC,GAAG,QAAQ,EAAE,QAAQ;AAC/C,eAAW,WAAW,kBAAkB;AACtC,UAAI,QAAQ,SAAS,UAAU;AAC7B,eAAO,QAAQ,OAAO;AACtB;AAAA,MACF,WAAW,QAAQ;AACjB;AAAA,MACF;AACA,UAAI,YAAY,KAAK,YAAY,OAAO;AACxC,UAAI,YAAY,WAAW;AACzB,iBAAS;AACT;AAAA,MACF;AACA,aAAO,QAAQ,OAAO;AACtB,mBAAa;AAAA,IACf;AAEA,WAAO;AAAA,EACT;AAAA,EAEc,UAAU,QAA2D;AAAA;AACjF,WAAK,0BAA0B,IAAI,wBAAwB;AAAA,QACzD,KAAK,KAAK;AAAA,MACZ,CAAC;AAED,WAAK,wBAAwB,GAAG,QAAQ,KAAK,MAAM;AACnD,WAAK,wBAAwB,GAAG,SAAS,KAAK,OAAO;AACrD,WAAK,wBAAwB,GAAG,OAAO,KAAK,KAAK;AAEjD,YAAM,KAAK,wBAAwB,MAAM,MAAM;AAAA,IACjD;AAAA;AAAA,EAoDQ,uBAAgC;AACtC,QAAI,OAAY;AAChB,QAAI;AACF,aAAO,KAAK,MAAM,KAAK,qBAAqB;AAAA,IAC9C,SAAS,OAAP;AACA,WAAK,KAAK,SAAS,KAAK;AACxB,WAAK,QAAQ;AACb,aAAO;AAAA,IACT;AACA,SAAK,KAAK,YAAY;AAAA,MACpB,MAAM,KAAK;AAAA,MACX,WAAW;AAAA,IACb,CAAC;AACD,SAAK,OAAO;AACZ,SAAK,mBAAmB;AACxB,SAAK,wBAAwB;AAC7B,WAAO;AAAA,EACT;AAAA,EAEQ,UAAU;AAChB,QAAI,KAAK,yBAAyB;AAChC,WAAK,wBAAwB,IAAI,QAAQ,KAAK,MAAM;AACpD,WAAK,wBAAwB,IAAI,SAAS,KAAK,OAAO;AACtD,WAAK,wBAAwB,IAAI,OAAO,KAAK,KAAK;AAAA,IACpD;AACA,SAAK,0BAA0B;AAC/B,SAAK,OAAO;AACZ,SAAK,mBAAmB;AACxB,SAAK,wBAAwB;AAAA,EAC/B;AAAA,EAEO,YAAY,SAA0B;AAC3C,QAAI,QAAQ,SAAS;AACnB,aAAO,eAAe,QAAQ,OAAO;AAAA,IACvC,WAAW,QAAQ,eAAe;AAChC,aAAO,eAAe,KAAK,UAAU,QAAQ,aAAa,CAAC;AAAA,IAC7D;AACA,WAAO;AAAA,EACT;AACF;AAEA,IAAM,mBAA8C;AAAA,EAClD,iBAAiB;AAAA,EACjB,qBAAqB;AAAA,EACrB,SAAS;AAAA,EACT,sBAAsB;AAAA,EACtB,aAAa;AAAA,EACb,sBAAsB;AAAA,EACtB,cAAc;AAAA,EACd,kBAAkB;AAAA,EAClB,sBAAsB;AAAA,EACtB,cAAc;AAAA,EACd,kBAAkB;AAAA,EAClB,0BAA0B;AAC5B;AAEA,SAAS,eAAe,MAAsB;AAC5C,SAAO,KAAK,SAAS;AACvB;AAEA,SAAS,qBAAqB,WAA+B;AAC3D,MAAI,UAAU,WAAW,GAAG;AAC1B,WAAO;AAAA,EACT;AACA,QAAM,OAAO,KAAK,UAAU,SAAS;AACrC,SAAO,eAAe,IAAI;AAC5B;;;AD7JO,SAAS,QAAQ,SAA0D;AAChF,QAAM,CAAC,UAAU,WAAW,IAAI,SAAoB,CAAC,CAAC;AACtD,QAAM,CAAC,OAAO,QAAQ,IAAI,SAAS,EAAE;AACrC,QAAM,CAAC,WAAW,YAAY,IAAI,SAAS,KAAK;AAEhD,QAAM,oBAAoB,CAAOC,cAA0C;AACzE,WAAO,IAAI,QAAiB,CAAC,SAAS,WAAW;AAC/C,mBAAa,IAAI;AAEjB,YAAM,mBAA4B;AAAA,QAChC,IAAI,OAAO;AAAA,QACX,WAAW,oBAAI,KAAK;AAAA,QACpB,SAAS;AAAA,QACT,MAAM;AAAA,MACR;AAGA,kBAAY,CAAC,GAAGA,WAAU,mBAAK,iBAAkB,CAAC;AAElD,YAAM,sBAAsB,CAAC,GAAI,QAAQ,mBAAmB,CAAC,GAAI,GAAGA,SAAQ;AAE5E,YAAM,SAAS,IAAI,qBAAqB;AAAA,QACtC,KAAK,QAAQ,OAAO;AAAA,MACtB,CAAC;AAED,YAAM,UAAU,MAAM;AACpB,eAAO,IAAI,SAAS;AACpB,eAAO,IAAI,KAAK;AAChB,eAAO,IAAI,OAAO;AAClB,eAAO,IAAI,UAAU;AAAA,MACvB;AAEA,aAAO,GAAG,WAAW,CAAC,YAAY;AAChC,yBAAiB,WAAW;AAC5B,oBAAY,CAAC,GAAGA,WAAU,mBAAK,iBAAkB,CAAC;AAAA,MACpD,CAAC;AAED,aAAO,GAAG,OAAO,MAAM;AACrB,qBAAa,KAAK;AAClB,gBAAQ;AACR,gBAAQ,mBAAK,iBAAkB;AAAA,MACjC,CAAC;AAED,aAAO,GAAG,SAAS,CAAC,UAAU;AAC5B,qBAAa,KAAK;AAClB,gBAAQ;AACR,eAAO,KAAK;AAAA,MACd,CAAC;AAED,aAAO,GAAG,YAAY,CAAO,iBAAiB;AAC5C,yBAAiB,gBAAgB;AAAA,UAC/B,MAAM,aAAa;AAAA,UACnB,WAAW,KAAK,UAAU,aAAa,SAAS;AAAA,QAClD;AACA,oBAAY,CAAC,GAAGA,WAAU,mBAAK,iBAAkB,CAAC;AAElD,qBAAa,KAAK;AAClB,gBAAQ;AACR,gBAAQ,mBAAK,iBAAkB;AAAA,MACjC,EAAC;AAED,aAAO,MAAM;AAAA,QACX,UAAU;AAAA,QACV,WAAW,QAAQ;AAAA,QACnB,SAAS,QAAQ;AAAA,QACjB,eAAe,QAAQ;AAAA,MACzB,CAAC;AAAA,IACH,CAAC;AAAA,EACH;AAEA,QAAM,yCAAyC,CAAOA,cAAuC;AAC3F,UAAM,UAAU,MAAM,kBAAkBA,SAAQ;AAChD,QAAI,QAAQ,iBAAiB,QAAQ,gBAAgB;AACnD,YAAM,QAAQ,eAAeA,WAAU,QAAQ,aAAa;AAAA,IAC9D;AAAA,EACF;AAEA,QAAM,SAAS,CAAO,YAAoC;AACxD,QAAI,WAAW;AACb;AAAA,IACF;AACA,UAAM,cAAc,CAAC,GAAG,UAAU,OAAO;AACzC,gBAAY,WAAW;AACvB,WAAO,uCAAuC,WAAW;AAAA,EAC3D;AAEA,QAAM,SAAS,MAA2B;AACxC,QAAI,aAAa,SAAS,WAAW,GAAG;AACtC;AAAA,IACF;AACA,QAAI,cAAc,CAAC,GAAG,QAAQ;AAC9B,UAAM,cAAc,SAAS,SAAS,SAAS,CAAC;AAEhD,QAAI,YAAY,SAAS,aAAa;AACpC,oBAAc,YAAY,MAAM,GAAG,EAAE;AAAA,IACvC;AACA,gBAAY,WAAW;AAEvB,WAAO,uCAAuC,WAAW;AAAA,EAC3D;AAEA,QAAM,OAAO,MAAY;AACvB,UAAM,IAAI,MAAM,iBAAiB;AAAA,EACnC;AAEA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,EACF;AACF;","names":["EventEmitter","EventEmitter","messages"]}
|
|
@@ -1,12 +1,19 @@
|
|
|
1
|
-
import {
|
|
1
|
+
import { Message } from '../types/base.js';
|
|
2
|
+
import { UseChatOptions } from './use-chat.js';
|
|
3
|
+
import '../context/copilot-context.js';
|
|
4
|
+
import '../types/annotated-function.js';
|
|
5
|
+
import '../types/document-pointer.js';
|
|
6
|
+
import 'react';
|
|
7
|
+
import './use-tree.js';
|
|
8
|
+
import 'openai/resources/chat';
|
|
2
9
|
|
|
3
10
|
interface UseCopilotChatOptions extends UseChatOptions {
|
|
4
11
|
makeSystemMessage?: (contextString: string) => string;
|
|
5
12
|
}
|
|
6
13
|
interface UseCopilotChatReturn {
|
|
7
14
|
visibleMessages: Message[];
|
|
8
|
-
append: (message: Message
|
|
9
|
-
reload: (
|
|
15
|
+
append: (message: Message) => Promise<void>;
|
|
16
|
+
reload: () => Promise<void>;
|
|
10
17
|
stop: () => void;
|
|
11
18
|
isLoading: boolean;
|
|
12
19
|
input: string;
|
|
@@ -1,11 +1,600 @@
|
|
|
1
|
-
|
|
2
|
-
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
|
|
1
|
+
var __defProp = Object.defineProperty;
|
|
2
|
+
var __defProps = Object.defineProperties;
|
|
3
|
+
var __getOwnPropDescs = Object.getOwnPropertyDescriptors;
|
|
4
|
+
var __getOwnPropSymbols = Object.getOwnPropertySymbols;
|
|
5
|
+
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
6
|
+
var __propIsEnum = Object.prototype.propertyIsEnumerable;
|
|
7
|
+
var __defNormalProp = (obj, key, value) => key in obj ? __defProp(obj, key, { enumerable: true, configurable: true, writable: true, value }) : obj[key] = value;
|
|
8
|
+
var __spreadValues = (a, b) => {
|
|
9
|
+
for (var prop in b || (b = {}))
|
|
10
|
+
if (__hasOwnProp.call(b, prop))
|
|
11
|
+
__defNormalProp(a, prop, b[prop]);
|
|
12
|
+
if (__getOwnPropSymbols)
|
|
13
|
+
for (var prop of __getOwnPropSymbols(b)) {
|
|
14
|
+
if (__propIsEnum.call(b, prop))
|
|
15
|
+
__defNormalProp(a, prop, b[prop]);
|
|
16
|
+
}
|
|
17
|
+
return a;
|
|
18
|
+
};
|
|
19
|
+
var __spreadProps = (a, b) => __defProps(a, __getOwnPropDescs(b));
|
|
20
|
+
var __objRest = (source, exclude) => {
|
|
21
|
+
var target = {};
|
|
22
|
+
for (var prop in source)
|
|
23
|
+
if (__hasOwnProp.call(source, prop) && exclude.indexOf(prop) < 0)
|
|
24
|
+
target[prop] = source[prop];
|
|
25
|
+
if (source != null && __getOwnPropSymbols)
|
|
26
|
+
for (var prop of __getOwnPropSymbols(source)) {
|
|
27
|
+
if (exclude.indexOf(prop) < 0 && __propIsEnum.call(source, prop))
|
|
28
|
+
target[prop] = source[prop];
|
|
29
|
+
}
|
|
30
|
+
return target;
|
|
31
|
+
};
|
|
32
|
+
var __async = (__this, __arguments, generator) => {
|
|
33
|
+
return new Promise((resolve, reject) => {
|
|
34
|
+
var fulfilled = (value) => {
|
|
35
|
+
try {
|
|
36
|
+
step(generator.next(value));
|
|
37
|
+
} catch (e) {
|
|
38
|
+
reject(e);
|
|
39
|
+
}
|
|
40
|
+
};
|
|
41
|
+
var rejected = (value) => {
|
|
42
|
+
try {
|
|
43
|
+
step(generator.throw(value));
|
|
44
|
+
} catch (e) {
|
|
45
|
+
reject(e);
|
|
46
|
+
}
|
|
47
|
+
};
|
|
48
|
+
var step = (x) => x.done ? resolve(x.value) : Promise.resolve(x.value).then(fulfilled, rejected);
|
|
49
|
+
step((generator = generator.apply(__this, __arguments)).next());
|
|
50
|
+
});
|
|
51
|
+
};
|
|
52
|
+
|
|
53
|
+
// src/hooks/use-copilot-chat.ts
|
|
54
|
+
import { useMemo, useContext } from "react";
|
|
55
|
+
|
|
56
|
+
// src/context/copilot-context.tsx
|
|
57
|
+
import React from "react";
|
|
58
|
+
var emptyCopilotContext = {
|
|
59
|
+
entryPoints: {},
|
|
60
|
+
setEntryPoint: () => {
|
|
61
|
+
},
|
|
62
|
+
removeEntryPoint: () => {
|
|
63
|
+
},
|
|
64
|
+
getChatCompletionFunctionDescriptions: () => returnAndThrowInDebug([]),
|
|
65
|
+
getFunctionCallHandler: () => returnAndThrowInDebug(() => __async(void 0, null, function* () {
|
|
66
|
+
})),
|
|
67
|
+
getContextString: (documents, categories) => returnAndThrowInDebug(""),
|
|
68
|
+
addContext: () => "",
|
|
69
|
+
removeContext: () => {
|
|
70
|
+
},
|
|
71
|
+
getDocumentsContext: (categories) => returnAndThrowInDebug([]),
|
|
72
|
+
addDocumentContext: () => returnAndThrowInDebug(""),
|
|
73
|
+
removeDocumentContext: () => {
|
|
74
|
+
},
|
|
75
|
+
copilotApiConfig: new class {
|
|
76
|
+
get chatApiEndpoint() {
|
|
77
|
+
throw new Error(
|
|
78
|
+
"Remember to wrap your app in a `<CopilotProvider> {...} </CopilotProvider>` !!!"
|
|
79
|
+
);
|
|
80
|
+
}
|
|
81
|
+
get chatApiEndpointV2() {
|
|
82
|
+
throw new Error(
|
|
83
|
+
"Remember to wrap your app in a `<CopilotProvider> {...} </CopilotProvider>` !!!"
|
|
84
|
+
);
|
|
85
|
+
}
|
|
86
|
+
get headers() {
|
|
87
|
+
return {};
|
|
88
|
+
}
|
|
89
|
+
get body() {
|
|
90
|
+
return {};
|
|
91
|
+
}
|
|
92
|
+
}()
|
|
93
|
+
};
|
|
94
|
+
var CopilotContext = React.createContext(emptyCopilotContext);
|
|
95
|
+
function returnAndThrowInDebug(value) {
|
|
96
|
+
throw new Error(
|
|
97
|
+
"Remember to wrap your app in a `<CopilotProvider> {...} </CopilotProvider>` !!!"
|
|
98
|
+
);
|
|
99
|
+
return value;
|
|
100
|
+
}
|
|
101
|
+
|
|
102
|
+
// src/hooks/use-chat.ts
|
|
103
|
+
import { useState } from "react";
|
|
104
|
+
import { nanoid } from "nanoid";
|
|
105
|
+
|
|
106
|
+
// src/openai/chat-completion-client.ts
|
|
107
|
+
import EventEmitter2 from "eventemitter3";
|
|
108
|
+
|
|
109
|
+
// src/openai/chat-completion-transport.ts
|
|
110
|
+
import EventEmitter from "eventemitter3";
|
|
111
|
+
var DEFAULT_MODEL = "gpt-4-1106-preview";
|
|
112
|
+
var ChatCompletionTransport = class extends EventEmitter {
|
|
113
|
+
constructor({ url }) {
|
|
114
|
+
super();
|
|
115
|
+
this.buffer = new Uint8Array();
|
|
116
|
+
this.bodyReader = null;
|
|
117
|
+
this.url = url;
|
|
118
|
+
}
|
|
119
|
+
cleanup() {
|
|
120
|
+
return __async(this, null, function* () {
|
|
121
|
+
if (this.bodyReader) {
|
|
122
|
+
try {
|
|
123
|
+
yield this.bodyReader.cancel();
|
|
124
|
+
} catch (error) {
|
|
125
|
+
console.warn("Failed to cancel body reader:", error);
|
|
126
|
+
}
|
|
127
|
+
}
|
|
128
|
+
this.bodyReader = null;
|
|
129
|
+
this.buffer = new Uint8Array();
|
|
130
|
+
});
|
|
131
|
+
}
|
|
132
|
+
fetch(_0) {
|
|
133
|
+
return __async(this, arguments, function* ({
|
|
134
|
+
model,
|
|
135
|
+
messages,
|
|
136
|
+
copilotConfig,
|
|
137
|
+
functions,
|
|
138
|
+
temperature,
|
|
139
|
+
headers,
|
|
140
|
+
body,
|
|
141
|
+
signal
|
|
142
|
+
}) {
|
|
143
|
+
yield this.cleanup();
|
|
144
|
+
temperature || (temperature = 0.5);
|
|
145
|
+
functions || (functions = []);
|
|
146
|
+
model || (model = DEFAULT_MODEL);
|
|
147
|
+
const cleanedMessages = messages.map((message) => {
|
|
148
|
+
const { content, role, name, function_call } = message;
|
|
149
|
+
return { content, role, name, function_call };
|
|
150
|
+
});
|
|
151
|
+
try {
|
|
152
|
+
const response = yield fetch(this.url, {
|
|
153
|
+
method: "POST",
|
|
154
|
+
headers: __spreadValues(__spreadValues({
|
|
155
|
+
"Content-Type": "application/json"
|
|
156
|
+
}, copilotConfig.headers), headers ? __spreadValues({}, headers) : {}),
|
|
157
|
+
body: JSON.stringify(__spreadValues(__spreadValues(__spreadValues(__spreadValues(__spreadValues({
|
|
158
|
+
model,
|
|
159
|
+
messages: cleanedMessages,
|
|
160
|
+
stream: true
|
|
161
|
+
}, functions.length ? { functions } : {}), temperature ? { temperature } : {}), functions.length != 0 ? { function_call: "auto" } : {}), copilotConfig.body), body ? __spreadValues({}, body) : {})),
|
|
162
|
+
signal
|
|
163
|
+
});
|
|
164
|
+
if (!response.ok) {
|
|
165
|
+
try {
|
|
166
|
+
const errorText = yield response.text();
|
|
167
|
+
yield this.cleanup();
|
|
168
|
+
const msg = `Error ${response.status}: ${errorText}`;
|
|
169
|
+
this.emit("error", new Error(msg));
|
|
170
|
+
} catch (_error) {
|
|
171
|
+
yield this.cleanup();
|
|
172
|
+
const msg = `Error ${response.status}: ${response.statusText}`;
|
|
173
|
+
this.emit("error", new Error(msg));
|
|
174
|
+
}
|
|
175
|
+
return;
|
|
176
|
+
}
|
|
177
|
+
if (response.body == null) {
|
|
178
|
+
yield this.cleanup();
|
|
179
|
+
const msg = "Response body is null";
|
|
180
|
+
this.emit("error", new Error(msg));
|
|
181
|
+
return;
|
|
182
|
+
}
|
|
183
|
+
this.bodyReader = response.body.getReader();
|
|
184
|
+
yield this.streamBody();
|
|
185
|
+
} catch (error) {
|
|
186
|
+
yield this.cleanup();
|
|
187
|
+
this.emit("error", error);
|
|
188
|
+
return;
|
|
189
|
+
}
|
|
190
|
+
});
|
|
191
|
+
}
|
|
192
|
+
streamBody() {
|
|
193
|
+
return __async(this, null, function* () {
|
|
194
|
+
while (true) {
|
|
195
|
+
try {
|
|
196
|
+
const { done, value } = yield this.bodyReader.read();
|
|
197
|
+
if (done) {
|
|
198
|
+
yield this.cleanup();
|
|
199
|
+
this.emit("end");
|
|
200
|
+
return;
|
|
201
|
+
}
|
|
202
|
+
const shouldContinue = yield this.processData(value);
|
|
203
|
+
if (!shouldContinue) {
|
|
204
|
+
return;
|
|
205
|
+
}
|
|
206
|
+
} catch (error) {
|
|
207
|
+
yield this.cleanup();
|
|
208
|
+
this.emit("error", error);
|
|
209
|
+
return;
|
|
210
|
+
}
|
|
211
|
+
}
|
|
212
|
+
});
|
|
213
|
+
}
|
|
214
|
+
processData(data) {
|
|
215
|
+
return __async(this, null, function* () {
|
|
216
|
+
const newBuffer = new Uint8Array(this.buffer.length + data.length);
|
|
217
|
+
newBuffer.set(this.buffer);
|
|
218
|
+
newBuffer.set(data, this.buffer.length);
|
|
219
|
+
this.buffer = newBuffer;
|
|
220
|
+
const dataString = new TextDecoder("utf-8").decode(this.buffer);
|
|
221
|
+
let lines = dataString.split("\n").filter((line) => line.trim() !== "");
|
|
222
|
+
if (!dataString.endsWith("\n")) {
|
|
223
|
+
const lastLine = lines.pop() || "";
|
|
224
|
+
const remainingBytes = new TextEncoder().encode(lastLine);
|
|
225
|
+
this.buffer = new Uint8Array(remainingBytes);
|
|
226
|
+
} else {
|
|
227
|
+
this.buffer = new Uint8Array();
|
|
228
|
+
}
|
|
229
|
+
for (const line of lines) {
|
|
230
|
+
const cleanedLine = line.replace(/^data: /, "");
|
|
231
|
+
if (cleanedLine === "[DONE]") {
|
|
232
|
+
yield this.cleanup();
|
|
233
|
+
this.emit("end");
|
|
234
|
+
return false;
|
|
235
|
+
}
|
|
236
|
+
let json;
|
|
237
|
+
try {
|
|
238
|
+
json = JSON.parse(cleanedLine);
|
|
239
|
+
} catch (error) {
|
|
240
|
+
console.error("Failed to parse JSON:", error);
|
|
241
|
+
continue;
|
|
242
|
+
}
|
|
243
|
+
this.emit("data", json);
|
|
244
|
+
}
|
|
245
|
+
return true;
|
|
246
|
+
});
|
|
247
|
+
}
|
|
248
|
+
};
|
|
249
|
+
|
|
250
|
+
// src/openai/chat-completion-client.ts
|
|
251
|
+
var DEFAULT_MAX_TOKENS = 8192;
|
|
252
|
+
var ChatCompletionClient = class extends EventEmitter2 {
|
|
253
|
+
constructor(params) {
|
|
254
|
+
super();
|
|
255
|
+
this.chatCompletionTransport = null;
|
|
256
|
+
this.mode = null;
|
|
257
|
+
this.functionCallName = "";
|
|
258
|
+
this.functionCallArguments = "";
|
|
259
|
+
this.onData = (data) => {
|
|
260
|
+
if (this.mode === "function" && !data.choices[0].delta.function_call) {
|
|
261
|
+
const success = this.tryFlushFunctionCall();
|
|
262
|
+
if (!success) {
|
|
263
|
+
return;
|
|
264
|
+
}
|
|
265
|
+
}
|
|
266
|
+
this.mode = data.choices[0].delta.function_call ? "function" : "message";
|
|
267
|
+
if (this.mode === "message") {
|
|
268
|
+
if (data.choices[0].delta.content) {
|
|
269
|
+
this.emit("content", data.choices[0].delta.content);
|
|
270
|
+
}
|
|
271
|
+
return;
|
|
272
|
+
} else if (this.mode === "function") {
|
|
273
|
+
if (data.choices[0].delta.function_call.name) {
|
|
274
|
+
this.functionCallName = data.choices[0].delta.function_call.name;
|
|
275
|
+
}
|
|
276
|
+
if (data.choices[0].delta.function_call.arguments) {
|
|
277
|
+
this.functionCallArguments += data.choices[0].delta.function_call.arguments;
|
|
278
|
+
}
|
|
279
|
+
this.emit("partial", this.functionCallName, this.functionCallArguments);
|
|
280
|
+
return;
|
|
281
|
+
}
|
|
282
|
+
};
|
|
283
|
+
this.onError = (error) => {
|
|
284
|
+
this.emit("error", error);
|
|
285
|
+
this.cleanup();
|
|
286
|
+
};
|
|
287
|
+
this.onEnd = () => {
|
|
288
|
+
if (this.mode === "function") {
|
|
289
|
+
const success = this.tryFlushFunctionCall();
|
|
290
|
+
if (!success) {
|
|
291
|
+
return;
|
|
292
|
+
}
|
|
293
|
+
}
|
|
294
|
+
this.emit("end");
|
|
295
|
+
this.cleanup();
|
|
296
|
+
};
|
|
297
|
+
this.model = params.model;
|
|
298
|
+
this.url = params.url;
|
|
299
|
+
}
|
|
300
|
+
fetch(params) {
|
|
301
|
+
return __async(this, null, function* () {
|
|
302
|
+
params = __spreadValues({}, params);
|
|
303
|
+
if (this.model && this.model in maxTokensByModel) {
|
|
304
|
+
params.maxTokens || (params.maxTokens = maxTokensByModel[this.model]);
|
|
305
|
+
} else {
|
|
306
|
+
params.maxTokens || (params.maxTokens = DEFAULT_MAX_TOKENS);
|
|
307
|
+
}
|
|
308
|
+
params.functions || (params.functions = []);
|
|
309
|
+
params.model = this.model;
|
|
310
|
+
params.messages = this.buildPrompt(params);
|
|
311
|
+
return yield this.runPrompt(params);
|
|
312
|
+
});
|
|
313
|
+
}
|
|
314
|
+
buildPrompt(params) {
|
|
315
|
+
let maxTokens = params.maxTokens;
|
|
316
|
+
const messages = params.messages;
|
|
317
|
+
const functions = params.functions;
|
|
318
|
+
const functionsNumTokens = countFunctionsTokens(functions);
|
|
319
|
+
if (functionsNumTokens > maxTokens) {
|
|
320
|
+
throw new Error(`Too many tokens in function calls: ${functionsNumTokens} > ${maxTokens}`);
|
|
321
|
+
}
|
|
322
|
+
maxTokens -= functionsNumTokens;
|
|
323
|
+
for (const message of messages) {
|
|
324
|
+
if (message.role === "system") {
|
|
325
|
+
const numTokens = this.countTokens(message);
|
|
326
|
+
maxTokens -= numTokens;
|
|
327
|
+
if (maxTokens < 0) {
|
|
328
|
+
throw new Error("Not enough tokens for system message.");
|
|
329
|
+
}
|
|
330
|
+
}
|
|
331
|
+
}
|
|
332
|
+
const result = [];
|
|
333
|
+
let cutoff = false;
|
|
334
|
+
const reversedMessages = [...messages].reverse();
|
|
335
|
+
for (const message of reversedMessages) {
|
|
336
|
+
if (message.role === "system") {
|
|
337
|
+
result.unshift(message);
|
|
338
|
+
continue;
|
|
339
|
+
} else if (cutoff) {
|
|
340
|
+
continue;
|
|
341
|
+
}
|
|
342
|
+
let numTokens = this.countTokens(message);
|
|
343
|
+
if (maxTokens < numTokens) {
|
|
344
|
+
cutoff = true;
|
|
345
|
+
continue;
|
|
346
|
+
}
|
|
347
|
+
result.unshift(message);
|
|
348
|
+
maxTokens -= numTokens;
|
|
349
|
+
}
|
|
350
|
+
return result;
|
|
351
|
+
}
|
|
352
|
+
runPrompt(params) {
|
|
353
|
+
return __async(this, null, function* () {
|
|
354
|
+
this.chatCompletionTransport = new ChatCompletionTransport({
|
|
355
|
+
url: this.url
|
|
356
|
+
});
|
|
357
|
+
this.chatCompletionTransport.on("data", this.onData);
|
|
358
|
+
this.chatCompletionTransport.on("error", this.onError);
|
|
359
|
+
this.chatCompletionTransport.on("end", this.onEnd);
|
|
360
|
+
yield this.chatCompletionTransport.fetch(params);
|
|
361
|
+
});
|
|
362
|
+
}
|
|
363
|
+
tryFlushFunctionCall() {
|
|
364
|
+
let args = null;
|
|
365
|
+
try {
|
|
366
|
+
args = JSON.parse(this.functionCallArguments);
|
|
367
|
+
} catch (error) {
|
|
368
|
+
this.emit("error", error);
|
|
369
|
+
this.cleanup();
|
|
370
|
+
return false;
|
|
371
|
+
}
|
|
372
|
+
this.emit("function", {
|
|
373
|
+
name: this.functionCallName,
|
|
374
|
+
arguments: args
|
|
375
|
+
});
|
|
376
|
+
this.mode = null;
|
|
377
|
+
this.functionCallName = "";
|
|
378
|
+
this.functionCallArguments = "";
|
|
379
|
+
return true;
|
|
380
|
+
}
|
|
381
|
+
cleanup() {
|
|
382
|
+
if (this.chatCompletionTransport) {
|
|
383
|
+
this.chatCompletionTransport.off("data", this.onData);
|
|
384
|
+
this.chatCompletionTransport.off("error", this.onError);
|
|
385
|
+
this.chatCompletionTransport.off("end", this.onEnd);
|
|
386
|
+
}
|
|
387
|
+
this.chatCompletionTransport = null;
|
|
388
|
+
this.mode = null;
|
|
389
|
+
this.functionCallName = "";
|
|
390
|
+
this.functionCallArguments = "";
|
|
391
|
+
}
|
|
392
|
+
countTokens(message) {
|
|
393
|
+
if (message.content) {
|
|
394
|
+
return estimateTokens(message.content);
|
|
395
|
+
} else if (message.function_call) {
|
|
396
|
+
return estimateTokens(JSON.stringify(message.function_call));
|
|
397
|
+
}
|
|
398
|
+
return 0;
|
|
399
|
+
}
|
|
400
|
+
};
|
|
401
|
+
var maxTokensByModel = {
|
|
402
|
+
"gpt-3.5-turbo": 4097,
|
|
403
|
+
"gpt-3.5-turbo-16k": 16385,
|
|
404
|
+
"gpt-4": 8192,
|
|
405
|
+
"gpt-4-1106-preview": 8192,
|
|
406
|
+
"gpt-4-32k": 32768,
|
|
407
|
+
"gpt-3.5-turbo-0301": 4097,
|
|
408
|
+
"gpt-4-0314": 8192,
|
|
409
|
+
"gpt-4-32k-0314": 32768,
|
|
410
|
+
"gpt-3.5-turbo-0613": 4097,
|
|
411
|
+
"gpt-4-0613": 8192,
|
|
412
|
+
"gpt-4-32k-0613": 32768,
|
|
413
|
+
"gpt-3.5-turbo-16k-0613": 16385
|
|
414
|
+
};
|
|
415
|
+
function estimateTokens(text) {
|
|
416
|
+
return text.length / 3;
|
|
417
|
+
}
|
|
418
|
+
function countFunctionsTokens(functions) {
|
|
419
|
+
if (functions.length === 0) {
|
|
420
|
+
return 0;
|
|
421
|
+
}
|
|
422
|
+
const json = JSON.stringify(functions);
|
|
423
|
+
return estimateTokens(json);
|
|
424
|
+
}
|
|
425
|
+
|
|
426
|
+
// src/hooks/use-chat.ts
|
|
427
|
+
function useChat(options) {
|
|
428
|
+
const [messages, setMessages] = useState([]);
|
|
429
|
+
const [input, setInput] = useState("");
|
|
430
|
+
const [isLoading, setIsLoading] = useState(false);
|
|
431
|
+
const runChatCompletion = (messages2) => __async(this, null, function* () {
|
|
432
|
+
return new Promise((resolve, reject) => {
|
|
433
|
+
setIsLoading(true);
|
|
434
|
+
const assistantMessage = {
|
|
435
|
+
id: nanoid(),
|
|
436
|
+
createdAt: /* @__PURE__ */ new Date(),
|
|
437
|
+
content: "",
|
|
438
|
+
role: "assistant"
|
|
439
|
+
};
|
|
440
|
+
setMessages([...messages2, __spreadValues({}, assistantMessage)]);
|
|
441
|
+
const messagesWithContext = [...options.initialMessages || [], ...messages2];
|
|
442
|
+
const client = new ChatCompletionClient({
|
|
443
|
+
url: options.api || "/api/copilotkit/openai"
|
|
444
|
+
});
|
|
445
|
+
const cleanup = () => {
|
|
446
|
+
client.off("content");
|
|
447
|
+
client.off("end");
|
|
448
|
+
client.off("error");
|
|
449
|
+
client.off("function");
|
|
450
|
+
};
|
|
451
|
+
client.on("content", (content) => {
|
|
452
|
+
assistantMessage.content += content;
|
|
453
|
+
setMessages([...messages2, __spreadValues({}, assistantMessage)]);
|
|
454
|
+
});
|
|
455
|
+
client.on("end", () => {
|
|
456
|
+
setIsLoading(false);
|
|
457
|
+
cleanup();
|
|
458
|
+
resolve(__spreadValues({}, assistantMessage));
|
|
459
|
+
});
|
|
460
|
+
client.on("error", (error) => {
|
|
461
|
+
setIsLoading(false);
|
|
462
|
+
cleanup();
|
|
463
|
+
reject(error);
|
|
464
|
+
});
|
|
465
|
+
client.on("function", (functionCall) => __async(this, null, function* () {
|
|
466
|
+
assistantMessage.function_call = {
|
|
467
|
+
name: functionCall.name,
|
|
468
|
+
arguments: JSON.stringify(functionCall.arguments)
|
|
469
|
+
};
|
|
470
|
+
setMessages([...messages2, __spreadValues({}, assistantMessage)]);
|
|
471
|
+
setIsLoading(false);
|
|
472
|
+
cleanup();
|
|
473
|
+
resolve(__spreadValues({}, assistantMessage));
|
|
474
|
+
}));
|
|
475
|
+
client.fetch({
|
|
476
|
+
messages: messagesWithContext,
|
|
477
|
+
functions: options.functions,
|
|
478
|
+
headers: options.headers,
|
|
479
|
+
copilotConfig: options.copilotConfig
|
|
480
|
+
});
|
|
481
|
+
});
|
|
482
|
+
});
|
|
483
|
+
const runChatCompletionAndHandleFunctionCall = (messages2) => __async(this, null, function* () {
|
|
484
|
+
const message = yield runChatCompletion(messages2);
|
|
485
|
+
if (message.function_call && options.onFunctionCall) {
|
|
486
|
+
yield options.onFunctionCall(messages2, message.function_call);
|
|
487
|
+
}
|
|
488
|
+
});
|
|
489
|
+
const append = (message) => __async(this, null, function* () {
|
|
490
|
+
if (isLoading) {
|
|
491
|
+
return;
|
|
492
|
+
}
|
|
493
|
+
const newMessages = [...messages, message];
|
|
494
|
+
setMessages(newMessages);
|
|
495
|
+
return runChatCompletionAndHandleFunctionCall(newMessages);
|
|
496
|
+
});
|
|
497
|
+
const reload = () => __async(this, null, function* () {
|
|
498
|
+
if (isLoading || messages.length === 0) {
|
|
499
|
+
return;
|
|
500
|
+
}
|
|
501
|
+
let newMessages = [...messages];
|
|
502
|
+
const lastMessage = messages[messages.length - 1];
|
|
503
|
+
if (lastMessage.role === "assistant") {
|
|
504
|
+
newMessages = newMessages.slice(0, -1);
|
|
505
|
+
}
|
|
506
|
+
setMessages(newMessages);
|
|
507
|
+
return runChatCompletionAndHandleFunctionCall(newMessages);
|
|
508
|
+
});
|
|
509
|
+
const stop = () => {
|
|
510
|
+
throw new Error("Not implemented");
|
|
511
|
+
};
|
|
512
|
+
return {
|
|
513
|
+
messages,
|
|
514
|
+
append,
|
|
515
|
+
reload,
|
|
516
|
+
stop,
|
|
517
|
+
isLoading,
|
|
518
|
+
input,
|
|
519
|
+
setInput
|
|
520
|
+
};
|
|
521
|
+
}
|
|
522
|
+
|
|
523
|
+
// src/components/copilot-provider/copilot-provider.tsx
|
|
524
|
+
import { useCallback, useState as useState2 } from "react";
|
|
525
|
+
import { jsx } from "react/jsx-runtime";
|
|
526
|
+
var defaultCopilotContextCategories = ["global"];
|
|
527
|
+
|
|
528
|
+
// src/hooks/use-copilot-chat.ts
|
|
529
|
+
function useCopilotChat(_a) {
|
|
530
|
+
var _b = _a, {
|
|
531
|
+
makeSystemMessage
|
|
532
|
+
} = _b, options = __objRest(_b, [
|
|
533
|
+
"makeSystemMessage"
|
|
534
|
+
]);
|
|
535
|
+
const {
|
|
536
|
+
getContextString,
|
|
537
|
+
getChatCompletionFunctionDescriptions,
|
|
538
|
+
getFunctionCallHandler,
|
|
539
|
+
copilotApiConfig
|
|
540
|
+
} = useContext(CopilotContext);
|
|
541
|
+
const systemMessage = useMemo(() => {
|
|
542
|
+
const systemMessageMaker = makeSystemMessage || defaultSystemMessage;
|
|
543
|
+
const contextString = getContextString([], defaultCopilotContextCategories);
|
|
544
|
+
return {
|
|
545
|
+
id: "system",
|
|
546
|
+
content: systemMessageMaker(contextString),
|
|
547
|
+
role: "system"
|
|
548
|
+
};
|
|
549
|
+
}, [getContextString, makeSystemMessage]);
|
|
550
|
+
const functionDescriptions = useMemo(() => {
|
|
551
|
+
return getChatCompletionFunctionDescriptions();
|
|
552
|
+
}, [getChatCompletionFunctionDescriptions]);
|
|
553
|
+
const { messages, append, reload, stop, isLoading, input, setInput } = useChat(__spreadProps(__spreadValues({}, options), {
|
|
554
|
+
copilotConfig: copilotApiConfig,
|
|
555
|
+
id: options.id,
|
|
556
|
+
initialMessages: [systemMessage].concat(options.initialMessages || []),
|
|
557
|
+
functions: functionDescriptions,
|
|
558
|
+
onFunctionCall: getFunctionCallHandler(),
|
|
559
|
+
headers: __spreadValues({}, options.headers),
|
|
560
|
+
body: __spreadValues({}, options.body)
|
|
561
|
+
}));
|
|
562
|
+
const visibleMessages = messages.filter(
|
|
563
|
+
(message) => message.role === "user" || message.role === "assistant"
|
|
564
|
+
);
|
|
565
|
+
return {
|
|
566
|
+
visibleMessages,
|
|
567
|
+
append,
|
|
568
|
+
reload,
|
|
569
|
+
stop,
|
|
570
|
+
isLoading,
|
|
571
|
+
input,
|
|
572
|
+
setInput
|
|
573
|
+
};
|
|
574
|
+
}
|
|
575
|
+
function defaultSystemMessage(contextString) {
|
|
576
|
+
return `
|
|
577
|
+
Please act as an efficient, competent, conscientious, and industrious professional assistant.
|
|
578
|
+
|
|
579
|
+
Help the user achieve their goals, and you do so in a way that is as efficient as possible, without unnecessary fluff, but also without sacrificing professionalism.
|
|
580
|
+
Always be polite and respectful, and prefer brevity over verbosity.
|
|
581
|
+
|
|
582
|
+
The user has provided you with the following context:
|
|
583
|
+
\`\`\`
|
|
584
|
+
${contextString}
|
|
585
|
+
\`\`\`
|
|
586
|
+
|
|
587
|
+
They have also provided you with functions you can call to initiate actions on their behalf, or functions you can call to receive more information.
|
|
588
|
+
|
|
589
|
+
Please assist them as best you can.
|
|
590
|
+
|
|
591
|
+
You can ask them for clarifying questions if needed, but don't be annoying about it. If you can reasonably 'fill in the blanks' yourself, do so.
|
|
592
|
+
|
|
593
|
+
If you would like to call a function, call it without saying anything else.
|
|
594
|
+
`;
|
|
595
|
+
}
|
|
596
|
+
export {
|
|
597
|
+
defaultSystemMessage,
|
|
598
|
+
useCopilotChat
|
|
599
|
+
};
|
|
11
600
|
//# sourceMappingURL=use-copilot-chat.mjs.map
|