190proof 1.0.1 → 1.0.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.d.mts +3 -32
- package/dist/index.d.ts +3 -32
- package/dist/index.js +11 -17
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +11 -14
- package/dist/index.mjs.map +1 -1
- package/package.json +1 -1
package/dist/index.d.mts
CHANGED
|
@@ -13,14 +13,9 @@ declare enum GPTModel {
|
|
|
13
13
|
declare enum GroqModel {
|
|
14
14
|
LLAMA_3_70B_8192 = "llama3-70b-8192"
|
|
15
15
|
}
|
|
16
|
-
declare enum Role {
|
|
17
|
-
User = "user",
|
|
18
|
-
Assistant = "assistant",
|
|
19
|
-
System = "system"
|
|
20
|
-
}
|
|
21
16
|
interface GenericMessage {
|
|
22
|
-
role:
|
|
23
|
-
content: string;
|
|
17
|
+
role: "user" | "assistant" | "system";
|
|
18
|
+
content: string | AnthropicContentBlock[];
|
|
24
19
|
timestamp?: string;
|
|
25
20
|
files?: File[];
|
|
26
21
|
functionCalls?: FunctionCall[];
|
|
@@ -31,14 +26,6 @@ interface File {
|
|
|
31
26
|
url?: string;
|
|
32
27
|
data?: string;
|
|
33
28
|
}
|
|
34
|
-
interface OpenAIMessage {
|
|
35
|
-
role: "user" | "assistant" | "system";
|
|
36
|
-
content: string;
|
|
37
|
-
}
|
|
38
|
-
interface AnthropicAIMessage {
|
|
39
|
-
role: "user" | "assistant" | "system";
|
|
40
|
-
content: string | AnthropicContentBlock[];
|
|
41
|
-
}
|
|
42
29
|
type AnthropicContentBlock = AnthropicTextContentBlock | AnthropicImageContentBlock;
|
|
43
30
|
interface AnthropicTextContentBlock {
|
|
44
31
|
type: "text";
|
|
@@ -80,20 +67,6 @@ interface OpenAIConfig {
|
|
|
80
67
|
interface AnthropicAIConfig {
|
|
81
68
|
service: "anthropic" | "bedrock";
|
|
82
69
|
}
|
|
83
|
-
interface OpenAIPayload {
|
|
84
|
-
model: GPTModel;
|
|
85
|
-
messages: OpenAIMessage[];
|
|
86
|
-
functions?: any[];
|
|
87
|
-
function_call?: "none" | "auto" | {
|
|
88
|
-
name: string;
|
|
89
|
-
};
|
|
90
|
-
temperature?: number;
|
|
91
|
-
}
|
|
92
|
-
interface AnthropicAIPayload {
|
|
93
|
-
model: ClaudeModel;
|
|
94
|
-
messages: AnthropicAIMessage[];
|
|
95
|
-
functions?: any[];
|
|
96
|
-
}
|
|
97
70
|
interface GenericPayload {
|
|
98
71
|
model: GPTModel | ClaudeModel | GroqModel;
|
|
99
72
|
messages: GenericMessage[];
|
|
@@ -104,8 +77,6 @@ interface GenericPayload {
|
|
|
104
77
|
temperature?: number;
|
|
105
78
|
}
|
|
106
79
|
|
|
107
|
-
declare function callOpenAiWithRetries(identifier: string, openAiPayload: OpenAIPayload, openAiConfig?: OpenAIConfig, retries?: number, chunkTimeoutMs?: number): Promise<ParsedResponseMessage>;
|
|
108
|
-
declare function callAnthropicWithRetries(identifier: string, AiPayload: AnthropicAIPayload, AiConfig?: AnthropicAIConfig, attempts?: number): Promise<ParsedResponseMessage>;
|
|
109
80
|
declare function callWithRetries(identifier: string, aiPayload: GenericPayload, aiConfig?: OpenAIConfig | AnthropicAIConfig, retries?: number, chunkTimeoutMs?: number): Promise<ParsedResponseMessage>;
|
|
110
81
|
|
|
111
|
-
export { ClaudeModel, GPTModel, GroqModel,
|
|
82
|
+
export { ClaudeModel, GPTModel, GroqModel, type OpenAIConfig, callWithRetries };
|
package/dist/index.d.ts
CHANGED
|
@@ -13,14 +13,9 @@ declare enum GPTModel {
|
|
|
13
13
|
declare enum GroqModel {
|
|
14
14
|
LLAMA_3_70B_8192 = "llama3-70b-8192"
|
|
15
15
|
}
|
|
16
|
-
declare enum Role {
|
|
17
|
-
User = "user",
|
|
18
|
-
Assistant = "assistant",
|
|
19
|
-
System = "system"
|
|
20
|
-
}
|
|
21
16
|
interface GenericMessage {
|
|
22
|
-
role:
|
|
23
|
-
content: string;
|
|
17
|
+
role: "user" | "assistant" | "system";
|
|
18
|
+
content: string | AnthropicContentBlock[];
|
|
24
19
|
timestamp?: string;
|
|
25
20
|
files?: File[];
|
|
26
21
|
functionCalls?: FunctionCall[];
|
|
@@ -31,14 +26,6 @@ interface File {
|
|
|
31
26
|
url?: string;
|
|
32
27
|
data?: string;
|
|
33
28
|
}
|
|
34
|
-
interface OpenAIMessage {
|
|
35
|
-
role: "user" | "assistant" | "system";
|
|
36
|
-
content: string;
|
|
37
|
-
}
|
|
38
|
-
interface AnthropicAIMessage {
|
|
39
|
-
role: "user" | "assistant" | "system";
|
|
40
|
-
content: string | AnthropicContentBlock[];
|
|
41
|
-
}
|
|
42
29
|
type AnthropicContentBlock = AnthropicTextContentBlock | AnthropicImageContentBlock;
|
|
43
30
|
interface AnthropicTextContentBlock {
|
|
44
31
|
type: "text";
|
|
@@ -80,20 +67,6 @@ interface OpenAIConfig {
|
|
|
80
67
|
interface AnthropicAIConfig {
|
|
81
68
|
service: "anthropic" | "bedrock";
|
|
82
69
|
}
|
|
83
|
-
interface OpenAIPayload {
|
|
84
|
-
model: GPTModel;
|
|
85
|
-
messages: OpenAIMessage[];
|
|
86
|
-
functions?: any[];
|
|
87
|
-
function_call?: "none" | "auto" | {
|
|
88
|
-
name: string;
|
|
89
|
-
};
|
|
90
|
-
temperature?: number;
|
|
91
|
-
}
|
|
92
|
-
interface AnthropicAIPayload {
|
|
93
|
-
model: ClaudeModel;
|
|
94
|
-
messages: AnthropicAIMessage[];
|
|
95
|
-
functions?: any[];
|
|
96
|
-
}
|
|
97
70
|
interface GenericPayload {
|
|
98
71
|
model: GPTModel | ClaudeModel | GroqModel;
|
|
99
72
|
messages: GenericMessage[];
|
|
@@ -104,8 +77,6 @@ interface GenericPayload {
|
|
|
104
77
|
temperature?: number;
|
|
105
78
|
}
|
|
106
79
|
|
|
107
|
-
declare function callOpenAiWithRetries(identifier: string, openAiPayload: OpenAIPayload, openAiConfig?: OpenAIConfig, retries?: number, chunkTimeoutMs?: number): Promise<ParsedResponseMessage>;
|
|
108
|
-
declare function callAnthropicWithRetries(identifier: string, AiPayload: AnthropicAIPayload, AiConfig?: AnthropicAIConfig, attempts?: number): Promise<ParsedResponseMessage>;
|
|
109
80
|
declare function callWithRetries(identifier: string, aiPayload: GenericPayload, aiConfig?: OpenAIConfig | AnthropicAIConfig, retries?: number, chunkTimeoutMs?: number): Promise<ParsedResponseMessage>;
|
|
110
81
|
|
|
111
|
-
export { ClaudeModel, GPTModel, GroqModel,
|
|
82
|
+
export { ClaudeModel, GPTModel, GroqModel, type OpenAIConfig, callWithRetries };
|
package/dist/index.js
CHANGED
|
@@ -27016,9 +27016,6 @@ __export(proof_exports, {
|
|
|
27016
27016
|
ClaudeModel: () => ClaudeModel,
|
|
27017
27017
|
GPTModel: () => GPTModel,
|
|
27018
27018
|
GroqModel: () => GroqModel,
|
|
27019
|
-
Role: () => Role,
|
|
27020
|
-
callAnthropicWithRetries: () => callAnthropicWithRetries,
|
|
27021
|
-
callOpenAiWithRetries: () => callOpenAiWithRetries,
|
|
27022
27019
|
callWithRetries: () => callWithRetries
|
|
27023
27020
|
});
|
|
27024
27021
|
module.exports = __toCommonJS(proof_exports);
|
|
@@ -27042,12 +27039,6 @@ var GroqModel = /* @__PURE__ */ ((GroqModel2) => {
|
|
|
27042
27039
|
GroqModel2["LLAMA_3_70B_8192"] = "llama3-70b-8192";
|
|
27043
27040
|
return GroqModel2;
|
|
27044
27041
|
})(GroqModel || {});
|
|
27045
|
-
var Role = /* @__PURE__ */ ((Role2) => {
|
|
27046
|
-
Role2["User"] = "user";
|
|
27047
|
-
Role2["Assistant"] = "assistant";
|
|
27048
|
-
Role2["System"] = "system";
|
|
27049
|
-
return Role2;
|
|
27050
|
-
})(Role || {});
|
|
27051
27042
|
|
|
27052
27043
|
// ../node_modules/@aws-sdk/client-bedrock-runtime/dist-es/BedrockRuntimeClient.js
|
|
27053
27044
|
init_dist_es3();
|
|
@@ -31532,10 +31523,14 @@ async function callOpenAiWithRetries(identifier, openAiPayload, openAiConfig, re
|
|
|
31532
31523
|
async function callOpenAIStream(identifier, openAiPayload, openAiConfig, chunkTimeoutMs) {
|
|
31533
31524
|
const functionNames = openAiPayload.functions ? new Set(openAiPayload.functions.map((fn) => fn.name)) : null;
|
|
31534
31525
|
if (!openAiConfig) {
|
|
31526
|
+
const defaultOpenAIBaseUrl = (
|
|
31527
|
+
// TODO: Remove this one we have per-provider configs
|
|
31528
|
+
"https://gateway.ai.cloudflare.com/v1/932636fc124abb5171fd630afe668905/igpt"
|
|
31529
|
+
);
|
|
31535
31530
|
openAiConfig = {
|
|
31536
31531
|
service: "openai",
|
|
31537
31532
|
apiKey: process.env.OPENAI_API_KEY,
|
|
31538
|
-
baseUrl:
|
|
31533
|
+
baseUrl: defaultOpenAIBaseUrl
|
|
31539
31534
|
};
|
|
31540
31535
|
}
|
|
31541
31536
|
let response;
|
|
@@ -31564,7 +31559,7 @@ async function callOpenAIStream(identifier, openAiPayload, openAiConfig, chunkTi
|
|
|
31564
31559
|
});
|
|
31565
31560
|
} else {
|
|
31566
31561
|
console.log(identifier, "Using OpenAI service", openAiPayload.model);
|
|
31567
|
-
const endpoint = `${openAiConfig
|
|
31562
|
+
const endpoint = `${openAiConfig.baseUrl}/openai/chat/completions`;
|
|
31568
31563
|
if (openAiConfig.orgId) {
|
|
31569
31564
|
console.log(identifier, "Using orgId", openAiConfig.orgId);
|
|
31570
31565
|
}
|
|
@@ -31765,7 +31760,6 @@ async function callAnthropic(identifier, AiPayload, AiConfig) {
|
|
|
31765
31760
|
data = response2.data;
|
|
31766
31761
|
}
|
|
31767
31762
|
const answers = data.content;
|
|
31768
|
-
console.log("Anthropic API answers:", JSON.stringify({ answers }));
|
|
31769
31763
|
if (!answers[0]) {
|
|
31770
31764
|
console.error(identifier, "Missing answer in Anthropic API:", data);
|
|
31771
31765
|
throw new Error("Missing answer in Anthropic API");
|
|
@@ -31911,7 +31905,7 @@ function prepareOpenAIPayload(payload) {
|
|
|
31911
31905
|
model: payload.model,
|
|
31912
31906
|
messages: payload.messages.map((message) => ({
|
|
31913
31907
|
role: message.role,
|
|
31914
|
-
content: message.content
|
|
31908
|
+
content: normalizeMessageContent(message.content)
|
|
31915
31909
|
// TODO: Handle files
|
|
31916
31910
|
})),
|
|
31917
31911
|
functions: payload.functions
|
|
@@ -31926,7 +31920,7 @@ function prepareGroqPayload(payload) {
|
|
|
31926
31920
|
model: payload.model,
|
|
31927
31921
|
messages: payload.messages.map((message) => ({
|
|
31928
31922
|
role: message.role,
|
|
31929
|
-
content: message.content
|
|
31923
|
+
content: normalizeMessageContent(message.content)
|
|
31930
31924
|
})),
|
|
31931
31925
|
functions: (_a3 = payload.functions) == null ? void 0 : _a3.map((fn) => ({
|
|
31932
31926
|
type: "function",
|
|
@@ -31934,6 +31928,9 @@ function prepareGroqPayload(payload) {
|
|
|
31934
31928
|
}))
|
|
31935
31929
|
};
|
|
31936
31930
|
}
|
|
31931
|
+
function normalizeMessageContent(content) {
|
|
31932
|
+
return Array.isArray(content) ? content.map((c5) => c5.type === "text" ? c5.text : `[${c5.type}]`).join("\n") : content;
|
|
31933
|
+
}
|
|
31937
31934
|
async function callGroq(identifier, payload) {
|
|
31938
31935
|
const response = await axios_default.post(
|
|
31939
31936
|
"https://api.groq.com/openai/v1/chat/completions",
|
|
@@ -31998,9 +31995,6 @@ async function callGroqWithRetries(identifier, payload, retries = 5) {
|
|
|
31998
31995
|
ClaudeModel,
|
|
31999
31996
|
GPTModel,
|
|
32000
31997
|
GroqModel,
|
|
32001
|
-
Role,
|
|
32002
|
-
callAnthropicWithRetries,
|
|
32003
|
-
callOpenAiWithRetries,
|
|
32004
31998
|
callWithRetries
|
|
32005
31999
|
});
|
|
32006
32000
|
/*! Bundled license information:
|