@copilotkit/react-core 0.16.0 → 0.17.0-alpha.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.turbo/turbo-build.log +168 -134
- package/CHANGELOG.md +6 -0
- package/dist/components/copilot-provider/copilot-provider-props.d.ts +3 -3
- package/dist/components/copilot-provider/copilot-provider-props.mjs +1 -2
- package/dist/components/copilot-provider/copilot-provider-props.mjs.map +1 -1
- package/dist/components/copilot-provider/copilot-provider.d.ts +3 -3
- package/dist/components/copilot-provider/copilot-provider.mjs +468 -7
- package/dist/components/copilot-provider/copilot-provider.mjs.map +1 -1
- package/dist/components/copilot-provider/index.d.ts +3 -3
- package/dist/components/copilot-provider/index.mjs +467 -8
- package/dist/components/copilot-provider/index.mjs.map +1 -1
- package/dist/components/copilot-provider/standard-copilot-api-config.d.ts +3 -3
- package/dist/components/copilot-provider/standard-copilot-api-config.mjs +12 -3
- package/dist/components/copilot-provider/standard-copilot-api-config.mjs.map +1 -1
- package/dist/components/index.d.ts +3 -3
- package/dist/components/index.mjs +467 -9
- package/dist/components/index.mjs.map +1 -1
- package/dist/context/copilot-context.d.ts +4 -8
- package/dist/context/copilot-context.mjs +70 -3
- package/dist/context/copilot-context.mjs.map +1 -1
- package/dist/context/index.d.ts +4 -4
- package/dist/context/index.mjs +69 -4
- package/dist/context/index.mjs.map +1 -1
- package/dist/hooks/index.d.ts +6 -1
- package/dist/hooks/index.mjs +654 -14
- package/dist/hooks/index.mjs.map +1 -1
- package/dist/hooks/use-chat.d.ts +84 -0
- package/dist/hooks/use-chat.mjs +461 -0
- package/dist/hooks/use-chat.mjs.map +1 -0
- package/dist/hooks/use-copilot-chat.d.ts +10 -3
- package/dist/hooks/use-copilot-chat.mjs +599 -10
- package/dist/hooks/use-copilot-chat.mjs.map +1 -1
- package/dist/hooks/use-flat-category-store.mjs +68 -3
- package/dist/hooks/use-flat-category-store.mjs.map +1 -1
- package/dist/hooks/use-make-copilot-actionable.mjs +95 -4
- package/dist/hooks/use-make-copilot-actionable.mjs.map +1 -1
- package/dist/hooks/use-make-copilot-document-readable.mjs +87 -4
- package/dist/hooks/use-make-copilot-document-readable.mjs.map +1 -1
- package/dist/hooks/use-make-copilot-readable.mjs +87 -4
- package/dist/hooks/use-make-copilot-readable.mjs.map +1 -1
- package/dist/hooks/use-tree.mjs +153 -3
- package/dist/hooks/use-tree.mjs.map +1 -1
- package/dist/index.d.ts +6 -2
- package/dist/index.mjs +1204 -21
- package/dist/index.mjs.map +1 -1
- package/dist/openai/chat-completion-client.d.ts +56 -0
- package/dist/openai/chat-completion-client.mjs +360 -0
- package/dist/openai/chat-completion-client.mjs.map +1 -0
- package/dist/openai/chat-completion-stream.d.ts +21 -0
- package/dist/openai/chat-completion-stream.mjs +221 -0
- package/dist/openai/chat-completion-stream.mjs.map +1 -0
- package/dist/openai/chat-completion-transport.d.ts +40 -0
- package/dist/openai/chat-completion-transport.mjs +181 -0
- package/dist/openai/chat-completion-transport.mjs.map +1 -0
- package/dist/openai/index.d.ts +10 -0
- package/dist/openai/index.mjs +221 -0
- package/dist/openai/index.mjs.map +1 -0
- package/dist/openai-assistants/hooks/index.mjs +235 -14
- package/dist/openai-assistants/hooks/index.mjs.map +1 -1
- package/dist/openai-assistants/hooks/use-assistants.mjs +52 -8
- package/dist/openai-assistants/hooks/use-assistants.mjs.map +1 -1
- package/dist/openai-assistants/hooks/use-copilot-chat-v2.mjs +236 -13
- package/dist/openai-assistants/hooks/use-copilot-chat-v2.mjs.map +1 -1
- package/dist/openai-assistants/index.mjs +236 -15
- package/dist/openai-assistants/index.mjs.map +1 -1
- package/dist/openai-assistants/utils/index.mjs +46 -4
- package/dist/openai-assistants/utils/index.mjs.map +1 -1
- package/dist/openai-assistants/utils/process-message-stream.mjs +46 -3
- package/dist/openai-assistants/utils/process-message-stream.mjs.map +1 -1
- package/dist/types/annotated-function.mjs +0 -2
- package/dist/types/annotated-function.mjs.map +1 -1
- package/dist/types/base.d.ts +56 -0
- package/dist/types/base.mjs +1 -0
- package/dist/types/base.mjs.map +1 -0
- package/dist/types/document-pointer.mjs +0 -2
- package/dist/types/document-pointer.mjs.map +1 -1
- package/dist/types/index.d.ts +1 -0
- package/dist/types/index.mjs +0 -2
- package/dist/types/index.mjs.map +1 -1
- package/dist/types/message.d.ts +2 -0
- package/dist/types/message.mjs +1 -0
- package/dist/types/message.mjs.map +1 -0
- package/dist/utils/utils.mjs +0 -2
- package/dist/utils/utils.mjs.map +1 -1
- package/dist/utils/utils.test.mjs +0 -1
- package/dist/utils/utils.test.mjs.map +1 -1
- package/package.json +4 -4
- package/src/components/copilot-provider/copilot-provider.tsx +4 -5
- package/src/context/copilot-context.tsx +1 -12
- package/src/context/index.ts +0 -1
- package/src/hooks/index.ts +1 -0
- package/src/hooks/use-chat.ts +197 -0
- package/src/hooks/use-copilot-chat.ts +10 -22
- package/src/index.tsx +1 -0
- package/src/openai/chat-completion-client.ts +240 -0
- package/src/openai/chat-completion-stream.ts +56 -0
- package/src/openai/chat-completion-transport.ts +190 -0
- package/src/openai/index.tsx +5 -0
- package/src/openai-assistants/hooks/use-copilot-chat-v2.ts +2 -2
- package/src/types/base.ts +61 -0
- package/src/types/index.ts +1 -0
- package/src/types/message.ts +0 -0
- package/dist/chunk-7GFKOIO7.mjs +0 -3
- package/dist/chunk-7GFKOIO7.mjs.map +0 -1
- package/dist/chunk-BABVSMJR.mjs +0 -3
- package/dist/chunk-BABVSMJR.mjs.map +0 -1
- package/dist/chunk-DE37LEZJ.mjs +0 -79
- package/dist/chunk-DE37LEZJ.mjs.map +0 -1
- package/dist/chunk-EFZPSZWO.mjs +0 -3
- package/dist/chunk-EFZPSZWO.mjs.map +0 -1
- package/dist/chunk-EV26IMLL.mjs +0 -165
- package/dist/chunk-EV26IMLL.mjs.map +0 -1
- package/dist/chunk-F2JIAPZQ.mjs +0 -13
- package/dist/chunk-F2JIAPZQ.mjs.map +0 -1
- package/dist/chunk-FRAKUJWH.mjs +0 -3
- package/dist/chunk-FRAKUJWH.mjs.map +0 -1
- package/dist/chunk-IF64NU27.mjs +0 -125
- package/dist/chunk-IF64NU27.mjs.map +0 -1
- package/dist/chunk-JD7BAH7U.mjs +0 -3
- package/dist/chunk-JD7BAH7U.mjs.map +0 -1
- package/dist/chunk-MRXNTQOX.mjs +0 -55
- package/dist/chunk-MRXNTQOX.mjs.map +0 -1
- package/dist/chunk-MZ5UN3BY.mjs +0 -28
- package/dist/chunk-MZ5UN3BY.mjs.map +0 -1
- package/dist/chunk-PF7LXYPO.mjs +0 -19
- package/dist/chunk-PF7LXYPO.mjs.map +0 -1
- package/dist/chunk-QACD2U6P.mjs +0 -3
- package/dist/chunk-QACD2U6P.mjs.map +0 -1
- package/dist/chunk-RFZQHCNS.mjs +0 -19
- package/dist/chunk-RFZQHCNS.mjs.map +0 -1
- package/dist/chunk-SPCZTZCY.mjs +0 -3
- package/dist/chunk-SPCZTZCY.mjs.map +0 -1
- package/dist/chunk-VUY2K2DI.mjs +0 -135
- package/dist/chunk-VUY2K2DI.mjs.map +0 -1
- package/dist/chunk-WL2MC3E2.mjs +0 -27
- package/dist/chunk-WL2MC3E2.mjs.map +0 -1
- package/dist/chunk-YGJFU4ZP.mjs +0 -58
- package/dist/chunk-YGJFU4ZP.mjs.map +0 -1
- package/dist/chunk-YPSGKPDA.mjs +0 -3
- package/dist/chunk-YPSGKPDA.mjs.map +0 -1
- package/dist/chunk-YULKJPY3.mjs +0 -70
- package/dist/chunk-YULKJPY3.mjs.map +0 -1
|
@@ -0,0 +1,21 @@
|
|
|
1
|
+
import { ChatCompletionTransportFetchParams } from './chat-completion-transport.js';
|
|
2
|
+
import 'eventemitter3';
|
|
3
|
+
import '../types/base.js';
|
|
4
|
+
import '../context/copilot-context.js';
|
|
5
|
+
import '../types/annotated-function.js';
|
|
6
|
+
import '../types/document-pointer.js';
|
|
7
|
+
import 'react';
|
|
8
|
+
import '../hooks/use-tree.js';
|
|
9
|
+
import 'openai/resources/chat';
|
|
10
|
+
|
|
11
|
+
interface ChatCompletionStreamConfiguration {
|
|
12
|
+
url: string;
|
|
13
|
+
model?: string;
|
|
14
|
+
}
|
|
15
|
+
declare class ChatCompletionStream {
|
|
16
|
+
private url;
|
|
17
|
+
constructor(params: ChatCompletionStreamConfiguration);
|
|
18
|
+
fetch(params: ChatCompletionTransportFetchParams): Promise<ReadableStream<string>>;
|
|
19
|
+
}
|
|
20
|
+
|
|
21
|
+
export { ChatCompletionStream, ChatCompletionStreamConfiguration };
|
|
@@ -0,0 +1,221 @@
|
|
|
1
|
+
var __defProp = Object.defineProperty;
|
|
2
|
+
var __getOwnPropSymbols = Object.getOwnPropertySymbols;
|
|
3
|
+
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
4
|
+
var __propIsEnum = Object.prototype.propertyIsEnumerable;
|
|
5
|
+
var __defNormalProp = (obj, key, value) => key in obj ? __defProp(obj, key, { enumerable: true, configurable: true, writable: true, value }) : obj[key] = value;
|
|
6
|
+
var __spreadValues = (a, b) => {
|
|
7
|
+
for (var prop in b || (b = {}))
|
|
8
|
+
if (__hasOwnProp.call(b, prop))
|
|
9
|
+
__defNormalProp(a, prop, b[prop]);
|
|
10
|
+
if (__getOwnPropSymbols)
|
|
11
|
+
for (var prop of __getOwnPropSymbols(b)) {
|
|
12
|
+
if (__propIsEnum.call(b, prop))
|
|
13
|
+
__defNormalProp(a, prop, b[prop]);
|
|
14
|
+
}
|
|
15
|
+
return a;
|
|
16
|
+
};
|
|
17
|
+
var __async = (__this, __arguments, generator) => {
|
|
18
|
+
return new Promise((resolve, reject) => {
|
|
19
|
+
var fulfilled = (value) => {
|
|
20
|
+
try {
|
|
21
|
+
step(generator.next(value));
|
|
22
|
+
} catch (e) {
|
|
23
|
+
reject(e);
|
|
24
|
+
}
|
|
25
|
+
};
|
|
26
|
+
var rejected = (value) => {
|
|
27
|
+
try {
|
|
28
|
+
step(generator.throw(value));
|
|
29
|
+
} catch (e) {
|
|
30
|
+
reject(e);
|
|
31
|
+
}
|
|
32
|
+
};
|
|
33
|
+
var step = (x) => x.done ? resolve(x.value) : Promise.resolve(x.value).then(fulfilled, rejected);
|
|
34
|
+
step((generator = generator.apply(__this, __arguments)).next());
|
|
35
|
+
});
|
|
36
|
+
};
|
|
37
|
+
|
|
38
|
+
// src/openai/chat-completion-transport.ts
|
|
39
|
+
import EventEmitter from "eventemitter3";
|
|
40
|
+
var DEFAULT_MODEL = "gpt-4-1106-preview";
|
|
41
|
+
var ChatCompletionTransport = class extends EventEmitter {
|
|
42
|
+
constructor({ url }) {
|
|
43
|
+
super();
|
|
44
|
+
this.buffer = new Uint8Array();
|
|
45
|
+
this.bodyReader = null;
|
|
46
|
+
this.url = url;
|
|
47
|
+
}
|
|
48
|
+
cleanup() {
|
|
49
|
+
return __async(this, null, function* () {
|
|
50
|
+
if (this.bodyReader) {
|
|
51
|
+
try {
|
|
52
|
+
yield this.bodyReader.cancel();
|
|
53
|
+
} catch (error) {
|
|
54
|
+
console.warn("Failed to cancel body reader:", error);
|
|
55
|
+
}
|
|
56
|
+
}
|
|
57
|
+
this.bodyReader = null;
|
|
58
|
+
this.buffer = new Uint8Array();
|
|
59
|
+
});
|
|
60
|
+
}
|
|
61
|
+
fetch(_0) {
|
|
62
|
+
return __async(this, arguments, function* ({
|
|
63
|
+
model,
|
|
64
|
+
messages,
|
|
65
|
+
copilotConfig,
|
|
66
|
+
functions,
|
|
67
|
+
temperature,
|
|
68
|
+
headers,
|
|
69
|
+
body,
|
|
70
|
+
signal
|
|
71
|
+
}) {
|
|
72
|
+
yield this.cleanup();
|
|
73
|
+
temperature || (temperature = 0.5);
|
|
74
|
+
functions || (functions = []);
|
|
75
|
+
model || (model = DEFAULT_MODEL);
|
|
76
|
+
const cleanedMessages = messages.map((message) => {
|
|
77
|
+
const { content, role, name, function_call } = message;
|
|
78
|
+
return { content, role, name, function_call };
|
|
79
|
+
});
|
|
80
|
+
try {
|
|
81
|
+
const response = yield fetch(this.url, {
|
|
82
|
+
method: "POST",
|
|
83
|
+
headers: __spreadValues(__spreadValues({
|
|
84
|
+
"Content-Type": "application/json"
|
|
85
|
+
}, copilotConfig.headers), headers ? __spreadValues({}, headers) : {}),
|
|
86
|
+
body: JSON.stringify(__spreadValues(__spreadValues(__spreadValues(__spreadValues(__spreadValues({
|
|
87
|
+
model,
|
|
88
|
+
messages: cleanedMessages,
|
|
89
|
+
stream: true
|
|
90
|
+
}, functions.length ? { functions } : {}), temperature ? { temperature } : {}), functions.length != 0 ? { function_call: "auto" } : {}), copilotConfig.body), body ? __spreadValues({}, body) : {})),
|
|
91
|
+
signal
|
|
92
|
+
});
|
|
93
|
+
if (!response.ok) {
|
|
94
|
+
try {
|
|
95
|
+
const errorText = yield response.text();
|
|
96
|
+
yield this.cleanup();
|
|
97
|
+
const msg = `Error ${response.status}: ${errorText}`;
|
|
98
|
+
this.emit("error", new Error(msg));
|
|
99
|
+
} catch (_error) {
|
|
100
|
+
yield this.cleanup();
|
|
101
|
+
const msg = `Error ${response.status}: ${response.statusText}`;
|
|
102
|
+
this.emit("error", new Error(msg));
|
|
103
|
+
}
|
|
104
|
+
return;
|
|
105
|
+
}
|
|
106
|
+
if (response.body == null) {
|
|
107
|
+
yield this.cleanup();
|
|
108
|
+
const msg = "Response body is null";
|
|
109
|
+
this.emit("error", new Error(msg));
|
|
110
|
+
return;
|
|
111
|
+
}
|
|
112
|
+
this.bodyReader = response.body.getReader();
|
|
113
|
+
yield this.streamBody();
|
|
114
|
+
} catch (error) {
|
|
115
|
+
yield this.cleanup();
|
|
116
|
+
this.emit("error", error);
|
|
117
|
+
return;
|
|
118
|
+
}
|
|
119
|
+
});
|
|
120
|
+
}
|
|
121
|
+
streamBody() {
|
|
122
|
+
return __async(this, null, function* () {
|
|
123
|
+
while (true) {
|
|
124
|
+
try {
|
|
125
|
+
const { done, value } = yield this.bodyReader.read();
|
|
126
|
+
if (done) {
|
|
127
|
+
yield this.cleanup();
|
|
128
|
+
this.emit("end");
|
|
129
|
+
return;
|
|
130
|
+
}
|
|
131
|
+
const shouldContinue = yield this.processData(value);
|
|
132
|
+
if (!shouldContinue) {
|
|
133
|
+
return;
|
|
134
|
+
}
|
|
135
|
+
} catch (error) {
|
|
136
|
+
yield this.cleanup();
|
|
137
|
+
this.emit("error", error);
|
|
138
|
+
return;
|
|
139
|
+
}
|
|
140
|
+
}
|
|
141
|
+
});
|
|
142
|
+
}
|
|
143
|
+
processData(data) {
|
|
144
|
+
return __async(this, null, function* () {
|
|
145
|
+
const newBuffer = new Uint8Array(this.buffer.length + data.length);
|
|
146
|
+
newBuffer.set(this.buffer);
|
|
147
|
+
newBuffer.set(data, this.buffer.length);
|
|
148
|
+
this.buffer = newBuffer;
|
|
149
|
+
const dataString = new TextDecoder("utf-8").decode(this.buffer);
|
|
150
|
+
let lines = dataString.split("\n").filter((line) => line.trim() !== "");
|
|
151
|
+
if (!dataString.endsWith("\n")) {
|
|
152
|
+
const lastLine = lines.pop() || "";
|
|
153
|
+
const remainingBytes = new TextEncoder().encode(lastLine);
|
|
154
|
+
this.buffer = new Uint8Array(remainingBytes);
|
|
155
|
+
} else {
|
|
156
|
+
this.buffer = new Uint8Array();
|
|
157
|
+
}
|
|
158
|
+
for (const line of lines) {
|
|
159
|
+
const cleanedLine = line.replace(/^data: /, "");
|
|
160
|
+
if (cleanedLine === "[DONE]") {
|
|
161
|
+
yield this.cleanup();
|
|
162
|
+
this.emit("end");
|
|
163
|
+
return false;
|
|
164
|
+
}
|
|
165
|
+
let json;
|
|
166
|
+
try {
|
|
167
|
+
json = JSON.parse(cleanedLine);
|
|
168
|
+
} catch (error) {
|
|
169
|
+
console.error("Failed to parse JSON:", error);
|
|
170
|
+
continue;
|
|
171
|
+
}
|
|
172
|
+
this.emit("data", json);
|
|
173
|
+
}
|
|
174
|
+
return true;
|
|
175
|
+
});
|
|
176
|
+
}
|
|
177
|
+
};
|
|
178
|
+
|
|
179
|
+
// src/openai/chat-completion-stream.ts
|
|
180
|
+
var ChatCompletionStream = class {
|
|
181
|
+
constructor(params) {
|
|
182
|
+
this.url = params.url;
|
|
183
|
+
}
|
|
184
|
+
fetch(params) {
|
|
185
|
+
return __async(this, null, function* () {
|
|
186
|
+
params = __spreadValues({}, params);
|
|
187
|
+
params.functions = void 0;
|
|
188
|
+
const transport = new ChatCompletionTransport({
|
|
189
|
+
url: this.url
|
|
190
|
+
});
|
|
191
|
+
const cleanup = () => {
|
|
192
|
+
transport.off("data");
|
|
193
|
+
transport.off("end");
|
|
194
|
+
transport.off("error");
|
|
195
|
+
};
|
|
196
|
+
const stream = new ReadableStream({
|
|
197
|
+
start: (controller) => {
|
|
198
|
+
transport.on("data", (data) => {
|
|
199
|
+
if (data.choices[0].delta.content) {
|
|
200
|
+
controller.enqueue(data.choices[0].delta.content);
|
|
201
|
+
}
|
|
202
|
+
});
|
|
203
|
+
transport.on("error", (error) => {
|
|
204
|
+
controller.error(error);
|
|
205
|
+
cleanup();
|
|
206
|
+
});
|
|
207
|
+
transport.on("end", () => {
|
|
208
|
+
controller.close();
|
|
209
|
+
cleanup();
|
|
210
|
+
});
|
|
211
|
+
}
|
|
212
|
+
});
|
|
213
|
+
transport.fetch(params);
|
|
214
|
+
return stream;
|
|
215
|
+
});
|
|
216
|
+
}
|
|
217
|
+
};
|
|
218
|
+
export {
|
|
219
|
+
ChatCompletionStream
|
|
220
|
+
};
|
|
221
|
+
//# sourceMappingURL=chat-completion-stream.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../../src/openai/chat-completion-transport.ts","../../src/openai/chat-completion-stream.ts"],"sourcesContent":["import EventEmitter from \"eventemitter3\";\nimport { Message, Function } from \"../types\";\nimport { CopilotApiConfig } from \"../context\";\n\nexport interface ChatCompletionTransportConfiguration {\n url: string;\n}\n\ninterface ChatCompletionTransportEvents {\n end: void;\n data: any;\n error: any;\n}\n\nexport interface ChatCompletionTransportFetchParams {\n model?: string;\n messages: Message[];\n functions?: Function[];\n temperature?: number;\n maxTokens?: number;\n headers?: Record<string, string> | Headers;\n body?: object;\n copilotConfig: CopilotApiConfig;\n signal?: AbortSignal;\n}\n\nconst DEFAULT_MODEL = \"gpt-4-1106-preview\";\n\nexport class ChatCompletionTransport extends EventEmitter<ChatCompletionTransportEvents> {\n private buffer = new Uint8Array();\n private bodyReader: ReadableStreamDefaultReader<Uint8Array> | null = null;\n private url: string;\n\n constructor({ url }: ChatCompletionTransportConfiguration) {\n super();\n this.url = url;\n }\n\n private async cleanup() {\n if (this.bodyReader) {\n try {\n await this.bodyReader.cancel();\n } catch (error) {\n console.warn(\"Failed to cancel body reader:\", error);\n }\n }\n this.bodyReader = null;\n this.buffer = new Uint8Array();\n }\n\n public async fetch({\n model,\n messages,\n copilotConfig,\n functions,\n temperature,\n headers,\n body,\n signal,\n }: ChatCompletionTransportFetchParams): Promise<void> {\n await this.cleanup();\n\n temperature ||= 0.5;\n functions ||= [];\n model ||= DEFAULT_MODEL;\n\n // clean up any extra properties from messages\n const cleanedMessages = messages.map((message) => {\n const { content, role, name, function_call } = message;\n return { content, role, name, function_call };\n });\n\n try {\n const response = await fetch(this.url, {\n method: \"POST\",\n headers: {\n \"Content-Type\": \"application/json\",\n ...copilotConfig.headers,\n ...(headers ? { ...headers } : {}),\n },\n body: JSON.stringify({\n model,\n messages: cleanedMessages,\n stream: true,\n ...(functions.length ? { functions } : {}),\n ...(temperature ? { temperature } : {}),\n ...(functions.length != 0 ? { function_call: \"auto\" } : {}),\n ...copilotConfig.body,\n ...(body ? { ...body } : {}),\n }),\n signal,\n });\n\n if (!response.ok) {\n try {\n const errorText = await response.text();\n await this.cleanup();\n const msg = `Error ${response.status}: ${errorText}`;\n this.emit(\"error\", new Error(msg));\n } catch (_error) {\n await this.cleanup();\n const msg = `Error ${response.status}: ${response.statusText}`;\n this.emit(\"error\", new Error(msg));\n }\n return;\n }\n\n if (response.body == null) {\n await this.cleanup();\n const msg = \"Response body is null\";\n this.emit(\"error\", new Error(msg));\n return;\n }\n\n this.bodyReader = response.body.getReader();\n\n await this.streamBody();\n } catch (error) {\n await this.cleanup();\n this.emit(\"error\", error);\n return;\n }\n }\n\n private async streamBody() {\n while (true) {\n try {\n const { done, value } = await this.bodyReader!.read();\n\n if (done) {\n await this.cleanup();\n this.emit(\"end\");\n return;\n }\n\n const shouldContinue = await this.processData(value);\n\n if (!shouldContinue) {\n return;\n }\n } catch (error) {\n await this.cleanup();\n this.emit(\"error\", error);\n return;\n }\n }\n }\n\n private async processData(data: Uint8Array): Promise<boolean> {\n // Append new data to the temp buffer\n const newBuffer = new Uint8Array(this.buffer.length + data.length);\n newBuffer.set(this.buffer);\n newBuffer.set(data, this.buffer.length);\n this.buffer = newBuffer;\n\n const dataString = new TextDecoder(\"utf-8\").decode(this.buffer);\n\n let lines = dataString.split(\"\\n\").filter((line) => line.trim() !== \"\");\n\n // If the last line isn't complete, keep it in the buffer for next time\n if (!dataString.endsWith(\"\\n\")) {\n const lastLine = lines.pop() || \"\";\n const remainingBytes = new TextEncoder().encode(lastLine);\n this.buffer = new Uint8Array(remainingBytes);\n } else {\n this.buffer = new Uint8Array();\n }\n\n for (const line of lines) {\n const cleanedLine = line.replace(/^data: /, \"\");\n\n if (cleanedLine === \"[DONE]\") {\n await this.cleanup();\n this.emit(\"end\");\n return false;\n }\n\n let json;\n try {\n json = JSON.parse(cleanedLine);\n } catch (error) {\n console.error(\"Failed to parse JSON:\", error);\n continue;\n }\n\n this.emit(\"data\", json);\n }\n return true;\n }\n}\n","import {\n ChatCompletionTransport,\n ChatCompletionTransportFetchParams,\n} from \"./chat-completion-transport\";\n\nexport interface ChatCompletionStreamConfiguration {\n url: string;\n model?: string;\n}\n\nexport class ChatCompletionStream {\n private url: string;\n\n constructor(params: ChatCompletionStreamConfiguration) {\n this.url = params.url;\n }\n\n public async fetch(params: ChatCompletionTransportFetchParams): Promise<ReadableStream<string>> {\n params = { ...params };\n params.functions = undefined;\n\n const transport = new ChatCompletionTransport({\n url: this.url,\n });\n\n const cleanup = () => {\n transport.off(\"data\");\n transport.off(\"end\");\n transport.off(\"error\");\n };\n\n const stream = new ReadableStream<string>({\n start: (controller) => {\n transport.on(\"data\", (data) => {\n if (data.choices[0].delta.content) {\n controller.enqueue(data.choices[0].delta.content);\n }\n });\n\n transport.on(\"error\", (error) => {\n controller.error(error);\n cleanup();\n });\n\n transport.on(\"end\", () => {\n controller.close();\n cleanup();\n });\n },\n });\n\n transport.fetch(params);\n\n return stream;\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA,OAAO,kBAAkB;AA0BzB,IAAM,gBAAgB;AAEf,IAAM,0BAAN,cAAsC,aAA4C;AAAA,EAKvF,YAAY,EAAE,IAAI,GAAyC;AACzD,UAAM;AALR,SAAQ,SAAS,IAAI,WAAW;AAChC,SAAQ,aAA6D;AAKnE,SAAK,MAAM;AAAA,EACb;AAAA,EAEc,UAAU;AAAA;AACtB,UAAI,KAAK,YAAY;AACnB,YAAI;AACF,gBAAM,KAAK,WAAW,OAAO;AAAA,QAC/B,SAAS,OAAP;AACA,kBAAQ,KAAK,iCAAiC,KAAK;AAAA,QACrD;AAAA,MACF;AACA,WAAK,aAAa;AAClB,WAAK,SAAS,IAAI,WAAW;AAAA,IAC/B;AAAA;AAAA,EAEa,MAAM,IASmC;AAAA,+CATnC;AAAA,MACjB;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF,GAAsD;AACpD,YAAM,KAAK,QAAQ;AAEnB,oCAAgB;AAChB,gCAAc,CAAC;AACf,wBAAU;AAGV,YAAM,kBAAkB,SAAS,IAAI,CAAC,YAAY;AAChD,cAAM,EAAE,SAAS,MAAM,MAAM,cAAc,IAAI;AAC/C,eAAO,EAAE,SAAS,MAAM,MAAM,cAAc;AAAA,MAC9C,CAAC;AAED,UAAI;AACF,cAAM,WAAW,MAAM,MAAM,KAAK,KAAK;AAAA,UACrC,QAAQ;AAAA,UACR,SAAS;AAAA,YACP,gBAAgB;AAAA,aACb,cAAc,UACb,UAAU,mBAAK,WAAY,CAAC;AAAA,UAElC,MAAM,KAAK,UAAU;AAAA,YACnB;AAAA,YACA,UAAU;AAAA,YACV,QAAQ;AAAA,aACJ,UAAU,SAAS,EAAE,UAAU,IAAI,CAAC,IACpC,cAAc,EAAE,YAAY,IAAI,CAAC,IACjC,UAAU,UAAU,IAAI,EAAE,eAAe,OAAO,IAAI,CAAC,IACtD,cAAc,OACb,OAAO,mBAAK,QAAS,CAAC,EAC3B;AAAA,UACD;AAAA,QACF,CAAC;AAED,YAAI,CAAC,SAAS,IAAI;AAChB,cAAI;AACF,kBAAM,YAAY,MAAM,SAAS,KAAK;AACtC,kBAAM,KAAK,QAAQ;AACnB,kBAAM,MAAM,SAAS,SAAS,WAAW;AACzC,iBAAK,KAAK,SAAS,IAAI,MAAM,GAAG,CAAC;AAAA,UACnC,SAAS,QAAP;AACA,kBAAM,KAAK,QAAQ;AACnB,kBAAM,MAAM,SAAS,SAAS,WAAW,SAAS;AAClD,iBAAK,KAAK,SAAS,IAAI,MAAM,GAAG,CAAC;AAAA,UACnC;AACA;AAAA,QACF;AAEA,YAAI,SAAS,QAAQ,MAAM;AACzB,gBAAM,KAAK,QAAQ;AACnB,gBAAM,MAAM;AACZ,eAAK,KAAK,SAAS,IAAI,MAAM,GAAG,CAAC;AACjC;AAAA,QACF;AAEA,aAAK,aAAa,SAAS,KAAK,UAAU;AAE1C,cAAM,KAAK,WAAW;AAAA,MACxB,SAAS,OAAP;AACA,cAAM,KAAK,QAAQ;AACnB,aAAK,KAAK,SAAS,KAAK;AACxB;AAAA,MACF;AAAA,IACF;AAAA;AAAA,EAEc,aAAa;AAAA;AACzB,aAAO,MAAM;AACX,YAAI;AACF,gBAAM,EAAE,MAAM,MAAM,IAAI,MAAM,KAAK,WAAY,KAAK;AAEpD,cAAI,MAAM;AACR,kBAAM,KAAK,QAAQ;AACnB,iBAAK,KAAK,KAAK;AACf;AAAA,UACF;AAEA,gBAAM,iBAAiB,MAAM,KAAK,YAAY,KAAK;AAEnD,cAAI,CAAC,gBAAgB;AACnB;AAAA,UACF;AAAA,QACF,SAAS,OAAP;AACA,gBAAM,KAAK,QAAQ;AACnB,eAAK,KAAK,SAAS,KAAK;AACxB;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAAA;AAAA,EAEc,YAAY,MAAoC;AAAA;AAE5D,YAAM,YAAY,IAAI,WAAW,KAAK,OAAO,SAAS,KAAK,MAAM;AACjE,gBAAU,IAAI,KAAK,MAAM;AACzB,gBAAU,IAAI,MAAM,KAAK,OAAO,MAAM;AACtC,WAAK,SAAS;AAEd,YAAM,aAAa,IAAI,YAAY,OAAO,EAAE,OAAO,KAAK,MAAM;AAE9D,UAAI,QAAQ,WAAW,MAAM,IAAI,EAAE,OAAO,CAAC,SAAS,KAAK,KAAK,MAAM,EAAE;AAGtE,UAAI,CAAC,WAAW,SAAS,IAAI,GAAG;AAC9B,cAAM,WAAW,MAAM,IAAI,KAAK;AAChC,cAAM,iBAAiB,IAAI,YAAY,EAAE,OAAO,QAAQ;AACxD,aAAK,SAAS,IAAI,WAAW,cAAc;AAAA,MAC7C,OAAO;AACL,aAAK,SAAS,IAAI,WAAW;AAAA,MAC/B;AAEA,iBAAW,QAAQ,OAAO;AACxB,cAAM,cAAc,KAAK,QAAQ,WAAW,EAAE;AAE9C,YAAI,gBAAgB,UAAU;AAC5B,gBAAM,KAAK,QAAQ;AACnB,eAAK,KAAK,KAAK;AACf,iBAAO;AAAA,QACT;AAEA,YAAI;AACJ,YAAI;AACF,iBAAO,KAAK,MAAM,WAAW;AAAA,QAC/B,SAAS,OAAP;AACA,kBAAQ,MAAM,yBAAyB,KAAK;AAC5C;AAAA,QACF;AAEA,aAAK,KAAK,QAAQ,IAAI;AAAA,MACxB;AACA,aAAO;AAAA,IACT;AAAA;AACF;;;ACnLO,IAAM,uBAAN,MAA2B;AAAA,EAGhC,YAAY,QAA2C;AACrD,SAAK,MAAM,OAAO;AAAA,EACpB;AAAA,EAEa,MAAM,QAA6E;AAAA;AAC9F,eAAS,mBAAK;AACd,aAAO,YAAY;AAEnB,YAAM,YAAY,IAAI,wBAAwB;AAAA,QAC5C,KAAK,KAAK;AAAA,MACZ,CAAC;AAED,YAAM,UAAU,MAAM;AACpB,kBAAU,IAAI,MAAM;AACpB,kBAAU,IAAI,KAAK;AACnB,kBAAU,IAAI,OAAO;AAAA,MACvB;AAEA,YAAM,SAAS,IAAI,eAAuB;AAAA,QACxC,OAAO,CAAC,eAAe;AACrB,oBAAU,GAAG,QAAQ,CAAC,SAAS;AAC7B,gBAAI,KAAK,QAAQ,CAAC,EAAE,MAAM,SAAS;AACjC,yBAAW,QAAQ,KAAK,QAAQ,CAAC,EAAE,MAAM,OAAO;AAAA,YAClD;AAAA,UACF,CAAC;AAED,oBAAU,GAAG,SAAS,CAAC,UAAU;AAC/B,uBAAW,MAAM,KAAK;AACtB,oBAAQ;AAAA,UACV,CAAC;AAED,oBAAU,GAAG,OAAO,MAAM;AACxB,uBAAW,MAAM;AACjB,oBAAQ;AAAA,UACV,CAAC;AAAA,QACH;AAAA,MACF,CAAC;AAED,gBAAU,MAAM,MAAM;AAEtB,aAAO;AAAA,IACT;AAAA;AACF;","names":[]}
|
|
@@ -0,0 +1,40 @@
|
|
|
1
|
+
import EventEmitter from 'eventemitter3';
|
|
2
|
+
import { Message, Function } from '../types/base.js';
|
|
3
|
+
import { CopilotApiConfig } from '../context/copilot-context.js';
|
|
4
|
+
import '../types/annotated-function.js';
|
|
5
|
+
import '../types/document-pointer.js';
|
|
6
|
+
import 'react';
|
|
7
|
+
import '../hooks/use-tree.js';
|
|
8
|
+
import 'openai/resources/chat';
|
|
9
|
+
|
|
10
|
+
interface ChatCompletionTransportConfiguration {
|
|
11
|
+
url: string;
|
|
12
|
+
}
|
|
13
|
+
interface ChatCompletionTransportEvents {
|
|
14
|
+
end: void;
|
|
15
|
+
data: any;
|
|
16
|
+
error: any;
|
|
17
|
+
}
|
|
18
|
+
interface ChatCompletionTransportFetchParams {
|
|
19
|
+
model?: string;
|
|
20
|
+
messages: Message[];
|
|
21
|
+
functions?: Function[];
|
|
22
|
+
temperature?: number;
|
|
23
|
+
maxTokens?: number;
|
|
24
|
+
headers?: Record<string, string> | Headers;
|
|
25
|
+
body?: object;
|
|
26
|
+
copilotConfig: CopilotApiConfig;
|
|
27
|
+
signal?: AbortSignal;
|
|
28
|
+
}
|
|
29
|
+
declare class ChatCompletionTransport extends EventEmitter<ChatCompletionTransportEvents> {
|
|
30
|
+
private buffer;
|
|
31
|
+
private bodyReader;
|
|
32
|
+
private url;
|
|
33
|
+
constructor({ url }: ChatCompletionTransportConfiguration);
|
|
34
|
+
private cleanup;
|
|
35
|
+
fetch({ model, messages, copilotConfig, functions, temperature, headers, body, signal, }: ChatCompletionTransportFetchParams): Promise<void>;
|
|
36
|
+
private streamBody;
|
|
37
|
+
private processData;
|
|
38
|
+
}
|
|
39
|
+
|
|
40
|
+
export { ChatCompletionTransport, ChatCompletionTransportConfiguration, ChatCompletionTransportFetchParams };
|
|
@@ -0,0 +1,181 @@
|
|
|
1
|
+
var __defProp = Object.defineProperty;
|
|
2
|
+
var __getOwnPropSymbols = Object.getOwnPropertySymbols;
|
|
3
|
+
var __hasOwnProp = Object.prototype.hasOwnProperty;
|
|
4
|
+
var __propIsEnum = Object.prototype.propertyIsEnumerable;
|
|
5
|
+
var __defNormalProp = (obj, key, value) => key in obj ? __defProp(obj, key, { enumerable: true, configurable: true, writable: true, value }) : obj[key] = value;
|
|
6
|
+
var __spreadValues = (a, b) => {
|
|
7
|
+
for (var prop in b || (b = {}))
|
|
8
|
+
if (__hasOwnProp.call(b, prop))
|
|
9
|
+
__defNormalProp(a, prop, b[prop]);
|
|
10
|
+
if (__getOwnPropSymbols)
|
|
11
|
+
for (var prop of __getOwnPropSymbols(b)) {
|
|
12
|
+
if (__propIsEnum.call(b, prop))
|
|
13
|
+
__defNormalProp(a, prop, b[prop]);
|
|
14
|
+
}
|
|
15
|
+
return a;
|
|
16
|
+
};
|
|
17
|
+
var __async = (__this, __arguments, generator) => {
|
|
18
|
+
return new Promise((resolve, reject) => {
|
|
19
|
+
var fulfilled = (value) => {
|
|
20
|
+
try {
|
|
21
|
+
step(generator.next(value));
|
|
22
|
+
} catch (e) {
|
|
23
|
+
reject(e);
|
|
24
|
+
}
|
|
25
|
+
};
|
|
26
|
+
var rejected = (value) => {
|
|
27
|
+
try {
|
|
28
|
+
step(generator.throw(value));
|
|
29
|
+
} catch (e) {
|
|
30
|
+
reject(e);
|
|
31
|
+
}
|
|
32
|
+
};
|
|
33
|
+
var step = (x) => x.done ? resolve(x.value) : Promise.resolve(x.value).then(fulfilled, rejected);
|
|
34
|
+
step((generator = generator.apply(__this, __arguments)).next());
|
|
35
|
+
});
|
|
36
|
+
};
|
|
37
|
+
|
|
38
|
+
// src/openai/chat-completion-transport.ts
|
|
39
|
+
import EventEmitter from "eventemitter3";
|
|
40
|
+
var DEFAULT_MODEL = "gpt-4-1106-preview";
|
|
41
|
+
var ChatCompletionTransport = class extends EventEmitter {
|
|
42
|
+
constructor({ url }) {
|
|
43
|
+
super();
|
|
44
|
+
this.buffer = new Uint8Array();
|
|
45
|
+
this.bodyReader = null;
|
|
46
|
+
this.url = url;
|
|
47
|
+
}
|
|
48
|
+
cleanup() {
|
|
49
|
+
return __async(this, null, function* () {
|
|
50
|
+
if (this.bodyReader) {
|
|
51
|
+
try {
|
|
52
|
+
yield this.bodyReader.cancel();
|
|
53
|
+
} catch (error) {
|
|
54
|
+
console.warn("Failed to cancel body reader:", error);
|
|
55
|
+
}
|
|
56
|
+
}
|
|
57
|
+
this.bodyReader = null;
|
|
58
|
+
this.buffer = new Uint8Array();
|
|
59
|
+
});
|
|
60
|
+
}
|
|
61
|
+
fetch(_0) {
|
|
62
|
+
return __async(this, arguments, function* ({
|
|
63
|
+
model,
|
|
64
|
+
messages,
|
|
65
|
+
copilotConfig,
|
|
66
|
+
functions,
|
|
67
|
+
temperature,
|
|
68
|
+
headers,
|
|
69
|
+
body,
|
|
70
|
+
signal
|
|
71
|
+
}) {
|
|
72
|
+
yield this.cleanup();
|
|
73
|
+
temperature || (temperature = 0.5);
|
|
74
|
+
functions || (functions = []);
|
|
75
|
+
model || (model = DEFAULT_MODEL);
|
|
76
|
+
const cleanedMessages = messages.map((message) => {
|
|
77
|
+
const { content, role, name, function_call } = message;
|
|
78
|
+
return { content, role, name, function_call };
|
|
79
|
+
});
|
|
80
|
+
try {
|
|
81
|
+
const response = yield fetch(this.url, {
|
|
82
|
+
method: "POST",
|
|
83
|
+
headers: __spreadValues(__spreadValues({
|
|
84
|
+
"Content-Type": "application/json"
|
|
85
|
+
}, copilotConfig.headers), headers ? __spreadValues({}, headers) : {}),
|
|
86
|
+
body: JSON.stringify(__spreadValues(__spreadValues(__spreadValues(__spreadValues(__spreadValues({
|
|
87
|
+
model,
|
|
88
|
+
messages: cleanedMessages,
|
|
89
|
+
stream: true
|
|
90
|
+
}, functions.length ? { functions } : {}), temperature ? { temperature } : {}), functions.length != 0 ? { function_call: "auto" } : {}), copilotConfig.body), body ? __spreadValues({}, body) : {})),
|
|
91
|
+
signal
|
|
92
|
+
});
|
|
93
|
+
if (!response.ok) {
|
|
94
|
+
try {
|
|
95
|
+
const errorText = yield response.text();
|
|
96
|
+
yield this.cleanup();
|
|
97
|
+
const msg = `Error ${response.status}: ${errorText}`;
|
|
98
|
+
this.emit("error", new Error(msg));
|
|
99
|
+
} catch (_error) {
|
|
100
|
+
yield this.cleanup();
|
|
101
|
+
const msg = `Error ${response.status}: ${response.statusText}`;
|
|
102
|
+
this.emit("error", new Error(msg));
|
|
103
|
+
}
|
|
104
|
+
return;
|
|
105
|
+
}
|
|
106
|
+
if (response.body == null) {
|
|
107
|
+
yield this.cleanup();
|
|
108
|
+
const msg = "Response body is null";
|
|
109
|
+
this.emit("error", new Error(msg));
|
|
110
|
+
return;
|
|
111
|
+
}
|
|
112
|
+
this.bodyReader = response.body.getReader();
|
|
113
|
+
yield this.streamBody();
|
|
114
|
+
} catch (error) {
|
|
115
|
+
yield this.cleanup();
|
|
116
|
+
this.emit("error", error);
|
|
117
|
+
return;
|
|
118
|
+
}
|
|
119
|
+
});
|
|
120
|
+
}
|
|
121
|
+
streamBody() {
|
|
122
|
+
return __async(this, null, function* () {
|
|
123
|
+
while (true) {
|
|
124
|
+
try {
|
|
125
|
+
const { done, value } = yield this.bodyReader.read();
|
|
126
|
+
if (done) {
|
|
127
|
+
yield this.cleanup();
|
|
128
|
+
this.emit("end");
|
|
129
|
+
return;
|
|
130
|
+
}
|
|
131
|
+
const shouldContinue = yield this.processData(value);
|
|
132
|
+
if (!shouldContinue) {
|
|
133
|
+
return;
|
|
134
|
+
}
|
|
135
|
+
} catch (error) {
|
|
136
|
+
yield this.cleanup();
|
|
137
|
+
this.emit("error", error);
|
|
138
|
+
return;
|
|
139
|
+
}
|
|
140
|
+
}
|
|
141
|
+
});
|
|
142
|
+
}
|
|
143
|
+
processData(data) {
|
|
144
|
+
return __async(this, null, function* () {
|
|
145
|
+
const newBuffer = new Uint8Array(this.buffer.length + data.length);
|
|
146
|
+
newBuffer.set(this.buffer);
|
|
147
|
+
newBuffer.set(data, this.buffer.length);
|
|
148
|
+
this.buffer = newBuffer;
|
|
149
|
+
const dataString = new TextDecoder("utf-8").decode(this.buffer);
|
|
150
|
+
let lines = dataString.split("\n").filter((line) => line.trim() !== "");
|
|
151
|
+
if (!dataString.endsWith("\n")) {
|
|
152
|
+
const lastLine = lines.pop() || "";
|
|
153
|
+
const remainingBytes = new TextEncoder().encode(lastLine);
|
|
154
|
+
this.buffer = new Uint8Array(remainingBytes);
|
|
155
|
+
} else {
|
|
156
|
+
this.buffer = new Uint8Array();
|
|
157
|
+
}
|
|
158
|
+
for (const line of lines) {
|
|
159
|
+
const cleanedLine = line.replace(/^data: /, "");
|
|
160
|
+
if (cleanedLine === "[DONE]") {
|
|
161
|
+
yield this.cleanup();
|
|
162
|
+
this.emit("end");
|
|
163
|
+
return false;
|
|
164
|
+
}
|
|
165
|
+
let json;
|
|
166
|
+
try {
|
|
167
|
+
json = JSON.parse(cleanedLine);
|
|
168
|
+
} catch (error) {
|
|
169
|
+
console.error("Failed to parse JSON:", error);
|
|
170
|
+
continue;
|
|
171
|
+
}
|
|
172
|
+
this.emit("data", json);
|
|
173
|
+
}
|
|
174
|
+
return true;
|
|
175
|
+
});
|
|
176
|
+
}
|
|
177
|
+
};
|
|
178
|
+
export {
|
|
179
|
+
ChatCompletionTransport
|
|
180
|
+
};
|
|
181
|
+
//# sourceMappingURL=chat-completion-transport.mjs.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"sources":["../../src/openai/chat-completion-transport.ts"],"sourcesContent":["import EventEmitter from \"eventemitter3\";\nimport { Message, Function } from \"../types\";\nimport { CopilotApiConfig } from \"../context\";\n\nexport interface ChatCompletionTransportConfiguration {\n url: string;\n}\n\ninterface ChatCompletionTransportEvents {\n end: void;\n data: any;\n error: any;\n}\n\nexport interface ChatCompletionTransportFetchParams {\n model?: string;\n messages: Message[];\n functions?: Function[];\n temperature?: number;\n maxTokens?: number;\n headers?: Record<string, string> | Headers;\n body?: object;\n copilotConfig: CopilotApiConfig;\n signal?: AbortSignal;\n}\n\nconst DEFAULT_MODEL = \"gpt-4-1106-preview\";\n\nexport class ChatCompletionTransport extends EventEmitter<ChatCompletionTransportEvents> {\n private buffer = new Uint8Array();\n private bodyReader: ReadableStreamDefaultReader<Uint8Array> | null = null;\n private url: string;\n\n constructor({ url }: ChatCompletionTransportConfiguration) {\n super();\n this.url = url;\n }\n\n private async cleanup() {\n if (this.bodyReader) {\n try {\n await this.bodyReader.cancel();\n } catch (error) {\n console.warn(\"Failed to cancel body reader:\", error);\n }\n }\n this.bodyReader = null;\n this.buffer = new Uint8Array();\n }\n\n public async fetch({\n model,\n messages,\n copilotConfig,\n functions,\n temperature,\n headers,\n body,\n signal,\n }: ChatCompletionTransportFetchParams): Promise<void> {\n await this.cleanup();\n\n temperature ||= 0.5;\n functions ||= [];\n model ||= DEFAULT_MODEL;\n\n // clean up any extra properties from messages\n const cleanedMessages = messages.map((message) => {\n const { content, role, name, function_call } = message;\n return { content, role, name, function_call };\n });\n\n try {\n const response = await fetch(this.url, {\n method: \"POST\",\n headers: {\n \"Content-Type\": \"application/json\",\n ...copilotConfig.headers,\n ...(headers ? { ...headers } : {}),\n },\n body: JSON.stringify({\n model,\n messages: cleanedMessages,\n stream: true,\n ...(functions.length ? { functions } : {}),\n ...(temperature ? { temperature } : {}),\n ...(functions.length != 0 ? { function_call: \"auto\" } : {}),\n ...copilotConfig.body,\n ...(body ? { ...body } : {}),\n }),\n signal,\n });\n\n if (!response.ok) {\n try {\n const errorText = await response.text();\n await this.cleanup();\n const msg = `Error ${response.status}: ${errorText}`;\n this.emit(\"error\", new Error(msg));\n } catch (_error) {\n await this.cleanup();\n const msg = `Error ${response.status}: ${response.statusText}`;\n this.emit(\"error\", new Error(msg));\n }\n return;\n }\n\n if (response.body == null) {\n await this.cleanup();\n const msg = \"Response body is null\";\n this.emit(\"error\", new Error(msg));\n return;\n }\n\n this.bodyReader = response.body.getReader();\n\n await this.streamBody();\n } catch (error) {\n await this.cleanup();\n this.emit(\"error\", error);\n return;\n }\n }\n\n private async streamBody() {\n while (true) {\n try {\n const { done, value } = await this.bodyReader!.read();\n\n if (done) {\n await this.cleanup();\n this.emit(\"end\");\n return;\n }\n\n const shouldContinue = await this.processData(value);\n\n if (!shouldContinue) {\n return;\n }\n } catch (error) {\n await this.cleanup();\n this.emit(\"error\", error);\n return;\n }\n }\n }\n\n private async processData(data: Uint8Array): Promise<boolean> {\n // Append new data to the temp buffer\n const newBuffer = new Uint8Array(this.buffer.length + data.length);\n newBuffer.set(this.buffer);\n newBuffer.set(data, this.buffer.length);\n this.buffer = newBuffer;\n\n const dataString = new TextDecoder(\"utf-8\").decode(this.buffer);\n\n let lines = dataString.split(\"\\n\").filter((line) => line.trim() !== \"\");\n\n // If the last line isn't complete, keep it in the buffer for next time\n if (!dataString.endsWith(\"\\n\")) {\n const lastLine = lines.pop() || \"\";\n const remainingBytes = new TextEncoder().encode(lastLine);\n this.buffer = new Uint8Array(remainingBytes);\n } else {\n this.buffer = new Uint8Array();\n }\n\n for (const line of lines) {\n const cleanedLine = line.replace(/^data: /, \"\");\n\n if (cleanedLine === \"[DONE]\") {\n await this.cleanup();\n this.emit(\"end\");\n return false;\n }\n\n let json;\n try {\n json = JSON.parse(cleanedLine);\n } catch (error) {\n console.error(\"Failed to parse JSON:\", error);\n continue;\n }\n\n this.emit(\"data\", json);\n }\n return true;\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA,OAAO,kBAAkB;AA0BzB,IAAM,gBAAgB;AAEf,IAAM,0BAAN,cAAsC,aAA4C;AAAA,EAKvF,YAAY,EAAE,IAAI,GAAyC;AACzD,UAAM;AALR,SAAQ,SAAS,IAAI,WAAW;AAChC,SAAQ,aAA6D;AAKnE,SAAK,MAAM;AAAA,EACb;AAAA,EAEc,UAAU;AAAA;AACtB,UAAI,KAAK,YAAY;AACnB,YAAI;AACF,gBAAM,KAAK,WAAW,OAAO;AAAA,QAC/B,SAAS,OAAP;AACA,kBAAQ,KAAK,iCAAiC,KAAK;AAAA,QACrD;AAAA,MACF;AACA,WAAK,aAAa;AAClB,WAAK,SAAS,IAAI,WAAW;AAAA,IAC/B;AAAA;AAAA,EAEa,MAAM,IASmC;AAAA,+CATnC;AAAA,MACjB;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF,GAAsD;AACpD,YAAM,KAAK,QAAQ;AAEnB,oCAAgB;AAChB,gCAAc,CAAC;AACf,wBAAU;AAGV,YAAM,kBAAkB,SAAS,IAAI,CAAC,YAAY;AAChD,cAAM,EAAE,SAAS,MAAM,MAAM,cAAc,IAAI;AAC/C,eAAO,EAAE,SAAS,MAAM,MAAM,cAAc;AAAA,MAC9C,CAAC;AAED,UAAI;AACF,cAAM,WAAW,MAAM,MAAM,KAAK,KAAK;AAAA,UACrC,QAAQ;AAAA,UACR,SAAS;AAAA,YACP,gBAAgB;AAAA,aACb,cAAc,UACb,UAAU,mBAAK,WAAY,CAAC;AAAA,UAElC,MAAM,KAAK,UAAU;AAAA,YACnB;AAAA,YACA,UAAU;AAAA,YACV,QAAQ;AAAA,aACJ,UAAU,SAAS,EAAE,UAAU,IAAI,CAAC,IACpC,cAAc,EAAE,YAAY,IAAI,CAAC,IACjC,UAAU,UAAU,IAAI,EAAE,eAAe,OAAO,IAAI,CAAC,IACtD,cAAc,OACb,OAAO,mBAAK,QAAS,CAAC,EAC3B;AAAA,UACD;AAAA,QACF,CAAC;AAED,YAAI,CAAC,SAAS,IAAI;AAChB,cAAI;AACF,kBAAM,YAAY,MAAM,SAAS,KAAK;AACtC,kBAAM,KAAK,QAAQ;AACnB,kBAAM,MAAM,SAAS,SAAS,WAAW;AACzC,iBAAK,KAAK,SAAS,IAAI,MAAM,GAAG,CAAC;AAAA,UACnC,SAAS,QAAP;AACA,kBAAM,KAAK,QAAQ;AACnB,kBAAM,MAAM,SAAS,SAAS,WAAW,SAAS;AAClD,iBAAK,KAAK,SAAS,IAAI,MAAM,GAAG,CAAC;AAAA,UACnC;AACA;AAAA,QACF;AAEA,YAAI,SAAS,QAAQ,MAAM;AACzB,gBAAM,KAAK,QAAQ;AACnB,gBAAM,MAAM;AACZ,eAAK,KAAK,SAAS,IAAI,MAAM,GAAG,CAAC;AACjC;AAAA,QACF;AAEA,aAAK,aAAa,SAAS,KAAK,UAAU;AAE1C,cAAM,KAAK,WAAW;AAAA,MACxB,SAAS,OAAP;AACA,cAAM,KAAK,QAAQ;AACnB,aAAK,KAAK,SAAS,KAAK;AACxB;AAAA,MACF;AAAA,IACF;AAAA;AAAA,EAEc,aAAa;AAAA;AACzB,aAAO,MAAM;AACX,YAAI;AACF,gBAAM,EAAE,MAAM,MAAM,IAAI,MAAM,KAAK,WAAY,KAAK;AAEpD,cAAI,MAAM;AACR,kBAAM,KAAK,QAAQ;AACnB,iBAAK,KAAK,KAAK;AACf;AAAA,UACF;AAEA,gBAAM,iBAAiB,MAAM,KAAK,YAAY,KAAK;AAEnD,cAAI,CAAC,gBAAgB;AACnB;AAAA,UACF;AAAA,QACF,SAAS,OAAP;AACA,gBAAM,KAAK,QAAQ;AACnB,eAAK,KAAK,SAAS,KAAK;AACxB;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAAA;AAAA,EAEc,YAAY,MAAoC;AAAA;AAE5D,YAAM,YAAY,IAAI,WAAW,KAAK,OAAO,SAAS,KAAK,MAAM;AACjE,gBAAU,IAAI,KAAK,MAAM;AACzB,gBAAU,IAAI,MAAM,KAAK,OAAO,MAAM;AACtC,WAAK,SAAS;AAEd,YAAM,aAAa,IAAI,YAAY,OAAO,EAAE,OAAO,KAAK,MAAM;AAE9D,UAAI,QAAQ,WAAW,MAAM,IAAI,EAAE,OAAO,CAAC,SAAS,KAAK,KAAK,MAAM,EAAE;AAGtE,UAAI,CAAC,WAAW,SAAS,IAAI,GAAG;AAC9B,cAAM,WAAW,MAAM,IAAI,KAAK;AAChC,cAAM,iBAAiB,IAAI,YAAY,EAAE,OAAO,QAAQ;AACxD,aAAK,SAAS,IAAI,WAAW,cAAc;AAAA,MAC7C,OAAO;AACL,aAAK,SAAS,IAAI,WAAW;AAAA,MAC/B;AAEA,iBAAW,QAAQ,OAAO;AACxB,cAAM,cAAc,KAAK,QAAQ,WAAW,EAAE;AAE9C,YAAI,gBAAgB,UAAU;AAC5B,gBAAM,KAAK,QAAQ;AACnB,eAAK,KAAK,KAAK;AACf,iBAAO;AAAA,QACT;AAEA,YAAI;AACJ,YAAI;AACF,iBAAO,KAAK,MAAM,WAAW;AAAA,QAC/B,SAAS,OAAP;AACA,kBAAQ,MAAM,yBAAyB,KAAK;AAC5C;AAAA,QACF;AAEA,aAAK,KAAK,QAAQ,IAAI;AAAA,MACxB;AACA,aAAO;AAAA,IACT;AAAA;AACF;","names":[]}
|
|
@@ -0,0 +1,10 @@
|
|
|
1
|
+
export { ChatCompletionStream, ChatCompletionStreamConfiguration } from './chat-completion-stream.js';
|
|
2
|
+
export { ChatCompletionTransportFetchParams } from './chat-completion-transport.js';
|
|
3
|
+
import 'eventemitter3';
|
|
4
|
+
import '../types/base.js';
|
|
5
|
+
import '../context/copilot-context.js';
|
|
6
|
+
import '../types/annotated-function.js';
|
|
7
|
+
import '../types/document-pointer.js';
|
|
8
|
+
import 'react';
|
|
9
|
+
import '../hooks/use-tree.js';
|
|
10
|
+
import 'openai/resources/chat';
|