langchain 0.0.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/cjs/agents/agent.d.ts +42 -0
- package/dist/cjs/agents/agent.js +129 -0
- package/dist/cjs/agents/agent.js.map +1 -0
- package/dist/cjs/agents/executor.d.ts +27 -0
- package/dist/cjs/agents/executor.js +105 -0
- package/dist/cjs/agents/executor.js.map +1 -0
- package/dist/cjs/agents/helpers.d.ts +3 -0
- package/dist/cjs/agents/helpers.js +30 -0
- package/dist/cjs/agents/helpers.js.map +1 -0
- package/dist/cjs/agents/index.d.ts +6 -0
- package/dist/cjs/agents/index.js +13 -0
- package/dist/cjs/agents/index.js.map +1 -0
- package/dist/cjs/agents/load.d.ts +6 -0
- package/dist/cjs/agents/load.js +28 -0
- package/dist/cjs/agents/load.js.map +1 -0
- package/dist/cjs/agents/mrkl/index.d.ts +33 -0
- package/dist/cjs/agents/mrkl/index.js +109 -0
- package/dist/cjs/agents/mrkl/index.js.map +1 -0
- package/dist/cjs/agents/mrkl/prompt.d.ts +3 -0
- package/dist/cjs/agents/mrkl/prompt.js +21 -0
- package/dist/cjs/agents/mrkl/prompt.js.map +1 -0
- package/dist/cjs/agents/tests/agent.test.d.ts +1 -0
- package/dist/cjs/agents/tests/agent.test.js +31 -0
- package/dist/cjs/agents/tests/agent.test.js.map +1 -0
- package/dist/cjs/agents/tools/index.d.ts +7 -0
- package/dist/cjs/agents/tools/index.js +6 -0
- package/dist/cjs/agents/tools/index.js.map +1 -0
- package/dist/cjs/agents/tools/llm-math.d.ts +0 -0
- package/dist/cjs/agents/tools/llm-math.js +2 -0
- package/dist/cjs/agents/tools/llm-math.js.map +1 -0
- package/dist/cjs/agents/tools/serpapi.d.ts +3 -0
- package/dist/cjs/agents/tools/serpapi.js +50 -0
- package/dist/cjs/agents/tools/serpapi.js.map +1 -0
- package/dist/cjs/agents/types.d.ts +25 -0
- package/dist/cjs/agents/types.js +3 -0
- package/dist/cjs/agents/types.js.map +1 -0
- package/dist/cjs/cache.d.ts +11 -0
- package/dist/cjs/cache.js +27 -0
- package/dist/cjs/cache.js.map +1 -0
- package/dist/cjs/chains/base.d.ts +12 -0
- package/dist/cjs/chains/base.js +23 -0
- package/dist/cjs/chains/base.js.map +1 -0
- package/dist/cjs/chains/index.d.ts +3 -0
- package/dist/cjs/chains/index.js +10 -0
- package/dist/cjs/chains/index.js.map +1 -0
- package/dist/cjs/chains/llm_chain.d.ts +30 -0
- package/dist/cjs/chains/llm_chain.js +83 -0
- package/dist/cjs/chains/llm_chain.js.map +1 -0
- package/dist/cjs/chains/llm_math_chain.d.ts +0 -0
- package/dist/cjs/chains/llm_math_chain.js +129 -0
- package/dist/cjs/chains/llm_math_chain.js.map +1 -0
- package/dist/cjs/chains/load.d.ts +2 -0
- package/dist/cjs/chains/load.js +29 -0
- package/dist/cjs/chains/load.js.map +1 -0
- package/dist/cjs/chains/tests/llm_chain.test.d.ts +1 -0
- package/dist/cjs/chains/tests/llm_chain.test.js +32 -0
- package/dist/cjs/chains/tests/llm_chain.test.js.map +1 -0
- package/dist/cjs/index.d.ts +3 -0
- package/dist/cjs/index.js +12 -0
- package/dist/cjs/index.js.map +1 -0
- package/dist/cjs/llms/base.d.ts +23 -0
- package/dist/cjs/llms/base.js +159 -0
- package/dist/cjs/llms/base.js.map +1 -0
- package/dist/cjs/llms/index.d.ts +18 -0
- package/dist/cjs/llms/index.js +11 -0
- package/dist/cjs/llms/index.js.map +1 -0
- package/dist/cjs/llms/load.d.ts +3 -0
- package/dist/cjs/llms/load.js +9 -0
- package/dist/cjs/llms/load.js.map +1 -0
- package/dist/cjs/llms/openai.d.ts +63 -0
- package/dist/cjs/llms/openai.js +202 -0
- package/dist/cjs/llms/openai.js.map +1 -0
- package/dist/cjs/llms/tests/openai.test.d.ts +1 -0
- package/dist/cjs/llms/tests/openai.test.js +19 -0
- package/dist/cjs/llms/tests/openai.test.js.map +1 -0
- package/dist/cjs/prompt/base.d.ts +19 -0
- package/dist/cjs/prompt/base.js +54 -0
- package/dist/cjs/prompt/base.js.map +1 -0
- package/dist/cjs/prompt/few_shot.d.ts +47 -0
- package/dist/cjs/prompt/few_shot.js +143 -0
- package/dist/cjs/prompt/few_shot.js.map +1 -0
- package/dist/cjs/prompt/index.d.ts +3 -0
- package/dist/cjs/prompt/index.js +10 -0
- package/dist/cjs/prompt/index.js.map +1 -0
- package/dist/cjs/prompt/load.d.ts +2 -0
- package/dist/cjs/prompt/load.js +25 -0
- package/dist/cjs/prompt/load.js.map +1 -0
- package/dist/cjs/prompt/parser.d.ts +40 -0
- package/dist/cjs/prompt/parser.js +93 -0
- package/dist/cjs/prompt/parser.js.map +1 -0
- package/dist/cjs/prompt/prompt.d.ts +28 -0
- package/dist/cjs/prompt/prompt.js +91 -0
- package/dist/cjs/prompt/prompt.js.map +1 -0
- package/dist/cjs/prompt/template.d.ts +16 -0
- package/dist/cjs/prompt/template.js +87 -0
- package/dist/cjs/prompt/template.js.map +1 -0
- package/dist/cjs/prompt/tests/load.test.d.ts +1 -0
- package/dist/cjs/prompt/tests/load.test.js +30 -0
- package/dist/cjs/prompt/tests/load.test.js.map +1 -0
- package/dist/cjs/prompt/tests/template.test.d.ts +1 -0
- package/dist/cjs/prompt/tests/template.test.js +27 -0
- package/dist/cjs/prompt/tests/template.test.js.map +1 -0
- package/dist/cjs/util/hub.d.ts +1 -0
- package/dist/cjs/util/hub.js +50 -0
- package/dist/cjs/util/hub.js.map +1 -0
- package/dist/cjs/util/index.d.ts +12 -0
- package/dist/cjs/util/index.js +107 -0
- package/dist/cjs/util/index.js.map +1 -0
- package/dist/esm/agents/agent.d.ts +42 -0
- package/dist/esm/agents/agent.js +124 -0
- package/dist/esm/agents/agent.js.map +1 -0
- package/dist/esm/agents/executor.d.ts +27 -0
- package/dist/esm/agents/executor.js +101 -0
- package/dist/esm/agents/executor.js.map +1 -0
- package/dist/esm/agents/helpers.d.ts +3 -0
- package/dist/esm/agents/helpers.js +26 -0
- package/dist/esm/agents/helpers.js.map +1 -0
- package/dist/esm/agents/index.d.ts +6 -0
- package/dist/esm/agents/index.js +5 -0
- package/dist/esm/agents/index.js.map +1 -0
- package/dist/esm/agents/load.d.ts +6 -0
- package/dist/esm/agents/load.js +24 -0
- package/dist/esm/agents/load.js.map +1 -0
- package/dist/esm/agents/mrkl/index.d.ts +33 -0
- package/dist/esm/agents/mrkl/index.js +106 -0
- package/dist/esm/agents/mrkl/index.js.map +1 -0
- package/dist/esm/agents/mrkl/prompt.d.ts +3 -0
- package/dist/esm/agents/mrkl/prompt.js +17 -0
- package/dist/esm/agents/mrkl/prompt.js.map +1 -0
- package/dist/esm/agents/tests/agent.test.d.ts +1 -0
- package/dist/esm/agents/tests/agent.test.js +29 -0
- package/dist/esm/agents/tests/agent.test.js.map +1 -0
- package/dist/esm/agents/tools/index.d.ts +7 -0
- package/dist/esm/agents/tools/index.js +2 -0
- package/dist/esm/agents/tools/index.js.map +1 -0
- package/dist/esm/agents/tools/llm-math.d.ts +0 -0
- package/dist/esm/agents/tools/llm-math.js +2 -0
- package/dist/esm/agents/tools/llm-math.js.map +1 -0
- package/dist/esm/agents/tools/serpapi.d.ts +3 -0
- package/dist/esm/agents/tools/serpapi.js +46 -0
- package/dist/esm/agents/tools/serpapi.js.map +1 -0
- package/dist/esm/agents/types.d.ts +25 -0
- package/dist/esm/agents/types.js +2 -0
- package/dist/esm/agents/types.js.map +1 -0
- package/dist/esm/cache.d.ts +11 -0
- package/dist/esm/cache.js +22 -0
- package/dist/esm/cache.js.map +1 -0
- package/dist/esm/chains/base.d.ts +12 -0
- package/dist/esm/chains/base.js +19 -0
- package/dist/esm/chains/base.js.map +1 -0
- package/dist/esm/chains/index.d.ts +3 -0
- package/dist/esm/chains/index.js +4 -0
- package/dist/esm/chains/index.js.map +1 -0
- package/dist/esm/chains/llm_chain.d.ts +30 -0
- package/dist/esm/chains/llm_chain.js +79 -0
- package/dist/esm/chains/llm_chain.js.map +1 -0
- package/dist/esm/chains/llm_math_chain.d.ts +0 -0
- package/dist/esm/chains/llm_math_chain.js +129 -0
- package/dist/esm/chains/llm_math_chain.js.map +1 -0
- package/dist/esm/chains/load.d.ts +2 -0
- package/dist/esm/chains/load.js +25 -0
- package/dist/esm/chains/load.js.map +1 -0
- package/dist/esm/chains/tests/llm_chain.test.d.ts +1 -0
- package/dist/esm/chains/tests/llm_chain.test.js +30 -0
- package/dist/esm/chains/tests/llm_chain.test.js.map +1 -0
- package/dist/esm/index.d.ts +3 -0
- package/dist/esm/index.js +4 -0
- package/dist/esm/index.js.map +1 -0
- package/dist/esm/llms/base.d.ts +23 -0
- package/dist/esm/llms/base.js +154 -0
- package/dist/esm/llms/base.js.map +1 -0
- package/dist/esm/llms/index.d.ts +18 -0
- package/dist/esm/llms/index.js +4 -0
- package/dist/esm/llms/index.js.map +1 -0
- package/dist/esm/llms/load.d.ts +3 -0
- package/dist/esm/llms/load.js +5 -0
- package/dist/esm/llms/load.js.map +1 -0
- package/dist/esm/llms/openai.d.ts +63 -0
- package/dist/esm/llms/openai.js +198 -0
- package/dist/esm/llms/openai.js.map +1 -0
- package/dist/esm/llms/tests/openai.test.d.ts +1 -0
- package/dist/esm/llms/tests/openai.test.js +17 -0
- package/dist/esm/llms/tests/openai.test.js.map +1 -0
- package/dist/esm/prompt/base.d.ts +19 -0
- package/dist/esm/prompt/base.js +50 -0
- package/dist/esm/prompt/base.js.map +1 -0
- package/dist/esm/prompt/few_shot.d.ts +47 -0
- package/dist/esm/prompt/few_shot.js +139 -0
- package/dist/esm/prompt/few_shot.js.map +1 -0
- package/dist/esm/prompt/index.d.ts +3 -0
- package/dist/esm/prompt/index.js +4 -0
- package/dist/esm/prompt/index.js.map +1 -0
- package/dist/esm/prompt/load.d.ts +2 -0
- package/dist/esm/prompt/load.js +21 -0
- package/dist/esm/prompt/load.js.map +1 -0
- package/dist/esm/prompt/parser.d.ts +40 -0
- package/dist/esm/prompt/parser.js +86 -0
- package/dist/esm/prompt/parser.js.map +1 -0
- package/dist/esm/prompt/prompt.d.ts +28 -0
- package/dist/esm/prompt/prompt.js +87 -0
- package/dist/esm/prompt/prompt.js.map +1 -0
- package/dist/esm/prompt/template.d.ts +16 -0
- package/dist/esm/prompt/template.js +80 -0
- package/dist/esm/prompt/template.js.map +1 -0
- package/dist/esm/prompt/tests/load.test.d.ts +1 -0
- package/dist/esm/prompt/tests/load.test.js +25 -0
- package/dist/esm/prompt/tests/load.test.js.map +1 -0
- package/dist/esm/prompt/tests/template.test.d.ts +1 -0
- package/dist/esm/prompt/tests/template.test.js +25 -0
- package/dist/esm/prompt/tests/template.test.js.map +1 -0
- package/dist/esm/util/hub.d.ts +1 -0
- package/dist/esm/util/hub.js +43 -0
- package/dist/esm/util/hub.js.map +1 -0
- package/dist/esm/util/index.d.ts +12 -0
- package/dist/esm/util/index.js +74 -0
- package/dist/esm/util/index.js.map +1 -0
- package/package.json +97 -0
|
@@ -0,0 +1,129 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
/*
|
|
3
|
+
import { BaseChain, ChainValues, LLMChain } from "./index";
|
|
4
|
+
|
|
5
|
+
import { BaseLLM, SerializedLLM } from "../llms";
|
|
6
|
+
import { BasePromptTemplate, SerializedBasePromptTemplate } from "../prompt";
|
|
7
|
+
|
|
8
|
+
import { resolveConfigFromFile } from "../util";
|
|
9
|
+
|
|
10
|
+
export interface LLMMathChainInput {
|
|
11
|
+
prompt: BasePromptTemplate;
|
|
12
|
+
llm: BaseLLM;
|
|
13
|
+
outputKey: string;
|
|
14
|
+
inputKey: string;
|
|
15
|
+
}
|
|
16
|
+
|
|
17
|
+
export type SerializedLLMChain = {
|
|
18
|
+
_type: "llm_chain";
|
|
19
|
+
llm?: SerializedLLM;
|
|
20
|
+
llm_path?: string;
|
|
21
|
+
prompt?: SerializedBasePromptTemplate;
|
|
22
|
+
prompt_path?: string;
|
|
23
|
+
};
|
|
24
|
+
|
|
25
|
+
export class LLMChain extends BaseChain implements LLMMathChainInput {
|
|
26
|
+
prompt: BasePromptTemplate;
|
|
27
|
+
|
|
28
|
+
llm: BaseLLM;
|
|
29
|
+
|
|
30
|
+
inputKey = "answer";
|
|
31
|
+
|
|
32
|
+
outputKey = "question";
|
|
33
|
+
|
|
34
|
+
constructor(fields: {
|
|
35
|
+
prompt: BasePromptTemplate;
|
|
36
|
+
llm: BaseLLM;
|
|
37
|
+
outputKey?: string;
|
|
38
|
+
inputKey?: string;
|
|
39
|
+
}) {
|
|
40
|
+
super();
|
|
41
|
+
this.prompt = fields.prompt;
|
|
42
|
+
this.llm = fields.llm;
|
|
43
|
+
this.outputKey = fields.outputKey ?? this.outputKey;
|
|
44
|
+
this.inputKey = fields.inputKey ?? this.inputKey;
|
|
45
|
+
}
|
|
46
|
+
|
|
47
|
+
async _call(values: ChainValues, stop?: string[]): Promise<ChainValues> {
|
|
48
|
+
const llm = new LLMChain({ prompt: this.prompt, llm: this.llm });
|
|
49
|
+
const p;
|
|
50
|
+
const formattedString = this.prompt.format(values);
|
|
51
|
+
const llmResult = await this.llm.call(formattedString, stop);
|
|
52
|
+
const result = { [this.outputKey]: llmResult };
|
|
53
|
+
return result;
|
|
54
|
+
}
|
|
55
|
+
|
|
56
|
+
async predict(values: ChainValues, stop?: string[]): Promise<string> {
|
|
57
|
+
const output = await this.call(values, stop);
|
|
58
|
+
return output[this.outputKey];
|
|
59
|
+
}
|
|
60
|
+
|
|
61
|
+
_chainType() {
|
|
62
|
+
return "llm_chain" as const;
|
|
63
|
+
}
|
|
64
|
+
|
|
65
|
+
static async deserialize(data: SerializedLLMChain) {
|
|
66
|
+
const serializedLLM = resolveConfigFromFile<"llm", SerializedLLM>(
|
|
67
|
+
"llm",
|
|
68
|
+
data
|
|
69
|
+
);
|
|
70
|
+
const serializedPrompt = resolveConfigFromFile<
|
|
71
|
+
"prompt",
|
|
72
|
+
SerializedBasePromptTemplate
|
|
73
|
+
>("prompt", data);
|
|
74
|
+
|
|
75
|
+
return new LLMChain({
|
|
76
|
+
llm: await BaseLLM.deserialize(serializedLLM),
|
|
77
|
+
prompt: await BasePromptTemplate.deserialize(serializedPrompt),
|
|
78
|
+
});
|
|
79
|
+
}
|
|
80
|
+
|
|
81
|
+
serialize(): SerializedLLMChain {
|
|
82
|
+
return {
|
|
83
|
+
_type: this._chainType(),
|
|
84
|
+
llm: this.llm.serialize(),
|
|
85
|
+
prompt: this.prompt.serialize(),
|
|
86
|
+
};
|
|
87
|
+
}
|
|
88
|
+
}
|
|
89
|
+
|
|
90
|
+
const PROMPT_TEMPLATE = `You are GPT-3, and you can't do math.
|
|
91
|
+
You can do basic math, and your memorization abilities are impressive, but you can't do any complex calculations that a human could not do in their head. You also have an annoying tendency to just make up highly specific, but wrong, answers.
|
|
92
|
+
|
|
93
|
+
So we hooked you up to a Python 3 kernel, and now you can execute code. If anyone gives you a hard math problem, just use this format and we’ll take care of the rest:
|
|
94
|
+
|
|
95
|
+
Question: \${{Question with hard calculation.}}
|
|
96
|
+
\`\`\`python
|
|
97
|
+
\${{Code that prints what you need to know}}
|
|
98
|
+
\`\`\`
|
|
99
|
+
\`\`\`output
|
|
100
|
+
\${{Output of your code}}
|
|
101
|
+
\`\`\`
|
|
102
|
+
Answer: \${{Answer}}
|
|
103
|
+
|
|
104
|
+
Otherwise, use this simpler format:
|
|
105
|
+
|
|
106
|
+
Question: \${{Question without hard calculation}}
|
|
107
|
+
Answer: \${{Answer}}
|
|
108
|
+
|
|
109
|
+
Begin.
|
|
110
|
+
|
|
111
|
+
Question: What is 37593 * 67?
|
|
112
|
+
|
|
113
|
+
\`\`\`python
|
|
114
|
+
print(37593 * 67)
|
|
115
|
+
\`\`\`
|
|
116
|
+
\`\`\`output
|
|
117
|
+
2518731
|
|
118
|
+
\`\`\`
|
|
119
|
+
Answer: 2518731
|
|
120
|
+
|
|
121
|
+
Question: {question}
|
|
122
|
+
`;
|
|
123
|
+
|
|
124
|
+
PROMPT = PromptTemplate(
|
|
125
|
+
(input_variables = ["question"]),
|
|
126
|
+
(template = _PROMPT_TEMPLATE)
|
|
127
|
+
);
|
|
128
|
+
*/
|
|
129
|
+
//# sourceMappingURL=llm_math_chain.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"llm_math_chain.js","sourceRoot":"","sources":["../../../chains/llm_math_chain.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EA8HE"}
|
|
@@ -0,0 +1,25 @@
|
|
|
1
|
+
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
|
2
|
+
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
|
3
|
+
return new (P || (P = Promise))(function (resolve, reject) {
|
|
4
|
+
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
|
5
|
+
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
|
6
|
+
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
|
7
|
+
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
|
8
|
+
});
|
|
9
|
+
};
|
|
10
|
+
import { BaseChain } from ".";
|
|
11
|
+
import { loadFromHub } from "../util/hub";
|
|
12
|
+
import { parseFileConfig } from "../util";
|
|
13
|
+
const loadChainFromFile = (file) => __awaiter(void 0, void 0, void 0, function* () {
|
|
14
|
+
const serialized = parseFileConfig(file);
|
|
15
|
+
console.log({ serialized });
|
|
16
|
+
return BaseChain.deserialize(serialized);
|
|
17
|
+
});
|
|
18
|
+
export const loadChain = (uri) => __awaiter(void 0, void 0, void 0, function* () {
|
|
19
|
+
const hubResult = yield loadFromHub(uri, loadChainFromFile, "chains", new Set(["json", "yaml"]));
|
|
20
|
+
if (hubResult) {
|
|
21
|
+
return hubResult;
|
|
22
|
+
}
|
|
23
|
+
return loadChainFromFile(uri);
|
|
24
|
+
});
|
|
25
|
+
//# sourceMappingURL=load.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"load.js","sourceRoot":"","sources":["../../../chains/load.ts"],"names":[],"mappings":";;;;;;;;;AAAA,OAAO,EAAE,SAAS,EAAE,MAAM,GAAG,CAAC;AAC9B,OAAO,EAAE,WAAW,EAAE,MAAM,aAAa,CAAC;AAC1C,OAAO,EAAE,eAAe,EAAE,MAAM,SAAS,CAAC;AAE1C,MAAM,iBAAiB,GAAG,CAAO,IAAY,EAAE,EAAE;IAC/C,MAAM,UAAU,GAAG,eAAe,CAAC,IAAI,CAAC,CAAC;IACzC,OAAO,CAAC,GAAG,CAAC,EAAE,UAAU,EAAE,CAAC,CAAC;IAC5B,OAAO,SAAS,CAAC,WAAW,CAAC,UAAU,CAAC,CAAC;AAC3C,CAAC,CAAA,CAAC;AAEF,MAAM,CAAC,MAAM,SAAS,GAAG,CAAO,GAAW,EAAsB,EAAE;IACjE,MAAM,SAAS,GAAG,MAAM,WAAW,CACjC,GAAG,EACH,iBAAiB,EACjB,QAAQ,EACR,IAAI,GAAG,CAAC,CAAC,MAAM,EAAE,MAAM,CAAC,CAAC,CAC1B,CAAC;IACF,IAAI,SAAS,EAAE;QACb,OAAO,SAAS,CAAC;KAClB;IAED,OAAO,iBAAiB,CAAC,GAAG,CAAC,CAAC;AAChC,CAAC,CAAA,CAAC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
export {};
|
|
@@ -0,0 +1,30 @@
|
|
|
1
|
+
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
|
2
|
+
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
|
3
|
+
return new (P || (P = Promise))(function (resolve, reject) {
|
|
4
|
+
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
|
5
|
+
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
|
6
|
+
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
|
7
|
+
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
|
8
|
+
});
|
|
9
|
+
};
|
|
10
|
+
import { test } from "@jest/globals";
|
|
11
|
+
import { OpenAI } from "../../llms/openai";
|
|
12
|
+
import { PromptTemplate } from "../../prompt";
|
|
13
|
+
import { LLMChain } from "../llm_chain";
|
|
14
|
+
import { loadChain } from "../load";
|
|
15
|
+
test("Test OpenAI", () => __awaiter(void 0, void 0, void 0, function* () {
|
|
16
|
+
const model = new OpenAI({});
|
|
17
|
+
const prompt = new PromptTemplate({
|
|
18
|
+
template: "Print {foo}",
|
|
19
|
+
inputVariables: ["foo"],
|
|
20
|
+
});
|
|
21
|
+
const chain = new LLMChain({ prompt, llm: model });
|
|
22
|
+
const res = yield chain.call({ foo: "my favorite color" });
|
|
23
|
+
console.log({ res });
|
|
24
|
+
}));
|
|
25
|
+
test("Load chain from hub", () => __awaiter(void 0, void 0, void 0, function* () {
|
|
26
|
+
const chain = yield loadChain("lc://chains/hello-world/chain.json");
|
|
27
|
+
const res = yield chain.call({ topic: "my favorite color" });
|
|
28
|
+
console.log({ res });
|
|
29
|
+
}));
|
|
30
|
+
//# sourceMappingURL=llm_chain.test.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"llm_chain.test.js","sourceRoot":"","sources":["../../../../chains/tests/llm_chain.test.ts"],"names":[],"mappings":";;;;;;;;;AAAA,OAAO,EAAE,IAAI,EAAE,MAAM,eAAe,CAAC;AACrC,OAAO,EAAE,MAAM,EAAE,MAAM,mBAAmB,CAAC;AAC3C,OAAO,EAAE,cAAc,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,EAAE,QAAQ,EAAE,MAAM,cAAc,CAAC;AACxC,OAAO,EAAE,SAAS,EAAE,MAAM,SAAS,CAAC;AAEpC,IAAI,CAAC,aAAa,EAAE,GAAS,EAAE;IAC7B,MAAM,KAAK,GAAG,IAAI,MAAM,CAAC,EAAE,CAAC,CAAC;IAC7B,MAAM,MAAM,GAAG,IAAI,cAAc,CAAC;QAChC,QAAQ,EAAE,aAAa;QACvB,cAAc,EAAE,CAAC,KAAK,CAAC;KACxB,CAAC,CAAC;IACH,MAAM,KAAK,GAAG,IAAI,QAAQ,CAAC,EAAE,MAAM,EAAE,GAAG,EAAE,KAAK,EAAE,CAAC,CAAC;IACnD,MAAM,GAAG,GAAG,MAAM,KAAK,CAAC,IAAI,CAAC,EAAE,GAAG,EAAE,mBAAmB,EAAE,CAAC,CAAC;IAC3D,OAAO,CAAC,GAAG,CAAC,EAAE,GAAG,EAAE,CAAC,CAAC;AACvB,CAAC,CAAA,CAAC,CAAC;AAEH,IAAI,CAAC,qBAAqB,EAAE,GAAS,EAAE;IACrC,MAAM,KAAK,GAAG,MAAM,SAAS,CAAC,oCAAoC,CAAC,CAAC;IACpE,MAAM,GAAG,GAAG,MAAM,KAAK,CAAC,IAAI,CAAC,EAAE,KAAK,EAAE,mBAAmB,EAAE,CAAC,CAAC;IAC7D,OAAO,CAAC,GAAG,CAAC,EAAE,GAAG,EAAE,CAAC,CAAC;AACvB,CAAC,CAAA,CAAC,CAAC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../index.ts"],"names":[],"mappings":"AAAA,OAAO,EACL,cAAc,EACd,kBAAkB,EAClB,qBAAqB,GACtB,MAAM,UAAU,CAAC;AAElB,OAAO,EAAE,QAAQ,EAAE,MAAM,UAAU,CAAC;AACpC,OAAO,EAAE,MAAM,EAAE,MAAM,QAAQ,CAAC"}
|
|
@@ -0,0 +1,23 @@
|
|
|
1
|
+
import { LLMCallbackManager, LLMResult } from "./index";
|
|
2
|
+
export type SerializedLLM = {
|
|
3
|
+
_type: string;
|
|
4
|
+
} & Record<string, any>;
|
|
5
|
+
export declare abstract class BaseLLM {
|
|
6
|
+
name: string;
|
|
7
|
+
cache?: boolean;
|
|
8
|
+
callbackManager: LLMCallbackManager;
|
|
9
|
+
verbose?: boolean;
|
|
10
|
+
constructor(callbackManager?: LLMCallbackManager, verbose?: boolean);
|
|
11
|
+
abstract _generate(prompts: string[], stop?: string[]): Promise<LLMResult>;
|
|
12
|
+
_generateUncached(prompts: string[], stop?: string[]): Promise<LLMResult>;
|
|
13
|
+
generate(prompts: string[], stop?: string[]): Promise<LLMResult>;
|
|
14
|
+
call(prompt: string, stop?: string[]): Promise<string>;
|
|
15
|
+
_identifyingParams(): Record<string, any>;
|
|
16
|
+
abstract _llmType(): string;
|
|
17
|
+
serialize(): SerializedLLM;
|
|
18
|
+
static deserialize(data: SerializedLLM): Promise<BaseLLM>;
|
|
19
|
+
}
|
|
20
|
+
export declare abstract class LLM extends BaseLLM {
|
|
21
|
+
abstract _call(prompt: string, stop?: string[]): Promise<string>;
|
|
22
|
+
_generate(prompts: string[], stop?: string[]): Promise<LLMResult>;
|
|
23
|
+
}
|
|
@@ -0,0 +1,154 @@
|
|
|
1
|
+
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
|
2
|
+
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
|
3
|
+
return new (P || (P = Promise))(function (resolve, reject) {
|
|
4
|
+
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
|
5
|
+
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
|
6
|
+
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
|
7
|
+
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
|
8
|
+
});
|
|
9
|
+
};
|
|
10
|
+
var __rest = (this && this.__rest) || function (s, e) {
|
|
11
|
+
var t = {};
|
|
12
|
+
for (var p in s) if (Object.prototype.hasOwnProperty.call(s, p) && e.indexOf(p) < 0)
|
|
13
|
+
t[p] = s[p];
|
|
14
|
+
if (s != null && typeof Object.getOwnPropertySymbols === "function")
|
|
15
|
+
for (var i = 0, p = Object.getOwnPropertySymbols(s); i < p.length; i++) {
|
|
16
|
+
if (e.indexOf(p[i]) < 0 && Object.prototype.propertyIsEnumerable.call(s, p[i]))
|
|
17
|
+
t[p[i]] = s[p[i]];
|
|
18
|
+
}
|
|
19
|
+
return t;
|
|
20
|
+
};
|
|
21
|
+
import { OpenAI } from "./index";
|
|
22
|
+
import { InMemoryCache } from "../cache";
|
|
23
|
+
const getCallbackManager = () => ({
|
|
24
|
+
handleStart: (..._args) => {
|
|
25
|
+
// console.log(args);
|
|
26
|
+
},
|
|
27
|
+
handleEnd: (..._args) => {
|
|
28
|
+
// console.log(args);
|
|
29
|
+
},
|
|
30
|
+
handleError: (..._args) => {
|
|
31
|
+
// console.log(args);
|
|
32
|
+
},
|
|
33
|
+
});
|
|
34
|
+
const getVerbosity = () => true;
|
|
35
|
+
const cache = new InMemoryCache();
|
|
36
|
+
export class BaseLLM {
|
|
37
|
+
constructor(callbackManager, verbose) {
|
|
38
|
+
Object.defineProperty(this, "name", {
|
|
39
|
+
enumerable: true,
|
|
40
|
+
configurable: true,
|
|
41
|
+
writable: true,
|
|
42
|
+
value: void 0
|
|
43
|
+
});
|
|
44
|
+
Object.defineProperty(this, "cache", {
|
|
45
|
+
enumerable: true,
|
|
46
|
+
configurable: true,
|
|
47
|
+
writable: true,
|
|
48
|
+
value: void 0
|
|
49
|
+
});
|
|
50
|
+
Object.defineProperty(this, "callbackManager", {
|
|
51
|
+
enumerable: true,
|
|
52
|
+
configurable: true,
|
|
53
|
+
writable: true,
|
|
54
|
+
value: void 0
|
|
55
|
+
});
|
|
56
|
+
Object.defineProperty(this, "verbose", {
|
|
57
|
+
enumerable: true,
|
|
58
|
+
configurable: true,
|
|
59
|
+
writable: true,
|
|
60
|
+
value: false
|
|
61
|
+
});
|
|
62
|
+
this.callbackManager = callbackManager !== null && callbackManager !== void 0 ? callbackManager : getCallbackManager();
|
|
63
|
+
this.verbose = verbose !== null && verbose !== void 0 ? verbose : getVerbosity();
|
|
64
|
+
}
|
|
65
|
+
_generateUncached(prompts, stop) {
|
|
66
|
+
return __awaiter(this, void 0, void 0, function* () {
|
|
67
|
+
this.callbackManager.handleStart({ name: this.name }, prompts, this.verbose);
|
|
68
|
+
let output;
|
|
69
|
+
try {
|
|
70
|
+
output = yield this._generate(prompts, stop);
|
|
71
|
+
}
|
|
72
|
+
catch (err) {
|
|
73
|
+
this.callbackManager.handleError(`${err}`, this.verbose);
|
|
74
|
+
throw err;
|
|
75
|
+
}
|
|
76
|
+
this.callbackManager.handleEnd(output, this.verbose);
|
|
77
|
+
return output;
|
|
78
|
+
});
|
|
79
|
+
}
|
|
80
|
+
generate(prompts, stop) {
|
|
81
|
+
var _a;
|
|
82
|
+
return __awaiter(this, void 0, void 0, function* () {
|
|
83
|
+
if (!Array.isArray(prompts)) {
|
|
84
|
+
throw new Error("Argument 'prompts' is expected to be a string[]");
|
|
85
|
+
}
|
|
86
|
+
if (this.cache === true && cache === null) {
|
|
87
|
+
throw new Error("Requested cache, but no cache found");
|
|
88
|
+
}
|
|
89
|
+
if (cache === null || this.cache === false) {
|
|
90
|
+
return this._generateUncached(prompts, stop);
|
|
91
|
+
}
|
|
92
|
+
const params = this.serialize();
|
|
93
|
+
params.stop = stop;
|
|
94
|
+
const llmStringKey = `${Object.entries(params).sort()}`;
|
|
95
|
+
const missingPromptIndices = [];
|
|
96
|
+
const generations = prompts.map((prompt, index) => {
|
|
97
|
+
const result = cache.lookup(prompt, llmStringKey);
|
|
98
|
+
if (!result) {
|
|
99
|
+
missingPromptIndices.push(index);
|
|
100
|
+
}
|
|
101
|
+
return result;
|
|
102
|
+
});
|
|
103
|
+
let llmOutput = {};
|
|
104
|
+
if (missingPromptIndices.length > 0) {
|
|
105
|
+
const results = yield this._generateUncached(missingPromptIndices.map((i) => prompts[i]), stop);
|
|
106
|
+
results.generations.forEach((generation, index) => {
|
|
107
|
+
const promptIndex = missingPromptIndices[index];
|
|
108
|
+
generations[promptIndex] = generation;
|
|
109
|
+
cache.update(prompts[promptIndex], llmStringKey, generation);
|
|
110
|
+
});
|
|
111
|
+
llmOutput = (_a = results.llmOutput) !== null && _a !== void 0 ? _a : {};
|
|
112
|
+
}
|
|
113
|
+
return { generations, llmOutput };
|
|
114
|
+
});
|
|
115
|
+
}
|
|
116
|
+
call(prompt, stop) {
|
|
117
|
+
return __awaiter(this, void 0, void 0, function* () {
|
|
118
|
+
const { generations } = yield this.generate([prompt], stop);
|
|
119
|
+
return generations[0][0].text;
|
|
120
|
+
});
|
|
121
|
+
}
|
|
122
|
+
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
123
|
+
_identifyingParams() {
|
|
124
|
+
return {};
|
|
125
|
+
}
|
|
126
|
+
serialize() {
|
|
127
|
+
return Object.assign(Object.assign({}, this._identifyingParams()), { _type: this._llmType() });
|
|
128
|
+
}
|
|
129
|
+
static deserialize(data) {
|
|
130
|
+
return __awaiter(this, void 0, void 0, function* () {
|
|
131
|
+
const { _type } = data, rest = __rest(data, ["_type"]);
|
|
132
|
+
const Cls = {
|
|
133
|
+
openai: OpenAI,
|
|
134
|
+
}[_type];
|
|
135
|
+
if (Cls === undefined) {
|
|
136
|
+
throw new Error(`Cannot load LLM with type ${_type}`);
|
|
137
|
+
}
|
|
138
|
+
return new Cls(rest);
|
|
139
|
+
});
|
|
140
|
+
}
|
|
141
|
+
}
|
|
142
|
+
export class LLM extends BaseLLM {
|
|
143
|
+
_generate(prompts, stop) {
|
|
144
|
+
return __awaiter(this, void 0, void 0, function* () {
|
|
145
|
+
const generations = [];
|
|
146
|
+
for (let i = 0; i < prompts.length; i += 1) {
|
|
147
|
+
const text = yield this._call(prompts[i], stop);
|
|
148
|
+
generations.push([{ text }]);
|
|
149
|
+
}
|
|
150
|
+
return { generations };
|
|
151
|
+
});
|
|
152
|
+
}
|
|
153
|
+
}
|
|
154
|
+
//# sourceMappingURL=base.js.map
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"base.js","sourceRoot":"","sources":["../../../llms/base.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA,OAAO,EAAiC,MAAM,EAAE,MAAM,SAAS,CAAC;AAChE,OAAO,EAAa,aAAa,EAAE,MAAM,UAAU,CAAC;AAEpD,MAAM,kBAAkB,GAAG,GAAuB,EAAE,CAAC,CAAC;IACpD,WAAW,EAAE,CAAC,GAAG,KAAK,EAAE,EAAE;QACxB,qBAAqB;IACvB,CAAC;IACD,SAAS,EAAE,CAAC,GAAG,KAAK,EAAE,EAAE;QACtB,qBAAqB;IACvB,CAAC;IACD,WAAW,EAAE,CAAC,GAAG,KAAK,EAAE,EAAE;QACxB,qBAAqB;IACvB,CAAC;CACF,CAAC,CAAC;AAEH,MAAM,YAAY,GAAG,GAAG,EAAE,CAAC,IAAI,CAAC;AAEhC,MAAM,KAAK,GAAc,IAAI,aAAa,EAAE,CAAC;AAO7C,MAAM,OAAgB,OAAO;IAS3B,YAAY,eAAoC,EAAE,OAAiB;QARnE;;;;;WAAa;QAEb;;;;;WAAgB;QAEhB;;;;;WAAoC;QAEpC;;;;mBAAoB,KAAK;WAAC;QAGxB,IAAI,CAAC,eAAe,GAAG,eAAe,aAAf,eAAe,cAAf,eAAe,GAAI,kBAAkB,EAAE,CAAC;QAC/D,IAAI,CAAC,OAAO,GAAG,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,YAAY,EAAE,CAAC;IAC3C,CAAC;IAIK,iBAAiB,CACrB,OAAiB,EACjB,IAAe;;YAEf,IAAI,CAAC,eAAe,CAAC,WAAW,CAC9B,EAAE,IAAI,EAAE,IAAI,CAAC,IAAI,EAAE,EACnB,OAAO,EACP,IAAI,CAAC,OAAO,CACb,CAAC;YACF,IAAI,MAAM,CAAC;YACX,IAAI;gBACF,MAAM,GAAG,MAAM,IAAI,CAAC,SAAS,CAAC,OAAO,EAAE,IAAI,CAAC,CAAC;aAC9C;YAAC,OAAO,GAAG,EAAE;gBACZ,IAAI,CAAC,eAAe,CAAC,WAAW,CAAC,GAAG,GAAG,EAAE,EAAE,IAAI,CAAC,OAAO,CAAC,CAAC;gBACzD,MAAM,GAAG,CAAC;aACX;YAED,IAAI,CAAC,eAAe,CAAC,SAAS,CAAC,MAAM,EAAE,IAAI,CAAC,OAAO,CAAC,CAAC;YACrD,OAAO,MAAM,CAAC;QAChB,CAAC;KAAA;IAEK,QAAQ,CAAC,OAAiB,EAAE,IAAe;;;YAC/C,IAAI,CAAC,KAAK,CAAC,OAAO,CAAC,OAAO,CAAC,EAAE;gBAC3B,MAAM,IAAI,KAAK,CAAC,iDAAiD,CAAC,CAAC;aACpE;YAED,IAAI,IAAI,CAAC,KAAK,KAAK,IAAI,IAAI,KAAK,KAAK,IAAI,EAAE;gBACzC,MAAM,IAAI,KAAK,CAAC,qCAAqC,CAAC,CAAC;aACxD;YAED,IAAI,KAAK,KAAK,IAAI,IAAI,IAAI,CAAC,KAAK,KAAK,KAAK,EAAE;gBAC1C,OAAO,IAAI,CAAC,iBAAiB,CAAC,OAAO,EAAE,IAAI,CAAC,CAAC;aAC9C;YAED,MAAM,MAAM,GAAG,IAAI,CAAC,SAAS,EAAE,CAAC;YAChC,MAAM,CAAC,IAAI,GAAG,IAAI,CAAC;YAEnB,MAAM,YAAY,GAAG,GAAG,MAAM,CAAC,OAAO,CAAC,MAAM,CAAC,CAAC,IAAI,EAAE,EAAE,CAAC;YACxD,MAAM,oBAAoB,GAAa,EAAE,CAAC;YAC1C,MAAM,WAAW,GAAG,OAAO,CAAC,GAAG,CAAC,CAAC,MAAM,EAAE,KAAK,EAAE,EAAE;gBAChD,MAAM,MAAM,GAAG,KAAK,CAAC,MAAM,CAAC,MAAM,EAAE,YAAY,CAAC,CAAC;gBAClD,IAAI,CAAC,MAAM,EAAE;oBACX,oBAAoB,CAAC,IAAI,CAAC,KAAK,CAAC,CAAC;iBAClC;gBACD,OAAO,MAAM,CAAC;YAChB,CAAC,CAAC,CAAC;YAEH,IAAI,SAAS,GAAG,EAAE,CAAC;YACnB,IAAI,oBAAoB,CAAC,MAAM,GAAG,CAAC,EAAE;gBACnC,MAAM,OAAO,GAAG,MAAM,IAAI,CAAC,iBAAiB,CAC1C,oBAAoB,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE,EAAE,CAAC,OAAO,CAAC,CAAC,CAAC,CAAC,EAC3C,IAAI,CACL,CAAC;gBACF,OAAO,CAAC,WAAW,CAAC,OAAO,CAAC,CAAC,UAAU,EAAE,KAAK,EAAE,EAAE;oBAChD,MAAM,WAAW,GAAG,oBAAoB,CAAC,KAAK,CAAC,CAAC;oBAChD,WAAW,CAAC,WAAW,CAAC,GAAG,UAAU,CAAC;oBACtC,KAAK,CAAC,MAAM,CAAC,OAAO,CAAC,WAAW,CAAC,EAAE,YAAY,EAAE,UAAU,CAAC,CAAC;gBAC/D,CAAC,CAAC,CAAC;gBACH,SAAS,GAAG,MAAA,OAAO,CAAC,SAAS,mCAAI,EAAE,CAAC;aACrC;YAED,OAAO,EAAE,WAAW,EAAE,SAAS,EAAe,CAAC;;KAChD;IAEK,IAAI,CAAC,MAAc,EAAE,IAAe;;YACxC,MAAM,EAAE,WAAW,EAAE,GAAG,MAAM,IAAI,CAAC,QAAQ,CAAC,CAAC,MAAM,CAAC,EAAE,IAAI,CAAC,CAAC;YAC5D,OAAO,WAAW,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,IAAI,CAAC;QAChC,CAAC;KAAA;IAED,8DAA8D;IAC9D,kBAAkB;QAChB,OAAO,EAAE,CAAC;IACZ,CAAC;IAID,SAAS;QACP,uCACK,IAAI,CAAC,kBAAkB,EAAE,KAC5B,KAAK,EAAE,IAAI,CAAC,QAAQ,EAAE,IACtB;IACJ,CAAC;IAED,MAAM,CAAO,WAAW,CAAC,IAAmB;;YAC1C,MAAM,EAAE,KAAK,KAAc,IAAI,EAAb,IAAI,UAAK,IAAI,EAAzB,SAAkB,CAAO,CAAC;YAChC,MAAM,GAAG,GAAG;gBACV,MAAM,EAAE,MAAM;aACf,CAAC,KAAK,CAAC,CAAC;YACT,IAAI,GAAG,KAAK,SAAS,EAAE;gBACrB,MAAM,IAAI,KAAK,CAAC,8BAA8B,KAAK,EAAE,CAAC,CAAC;aACxD;YACD,OAAO,IAAI,GAAG,CAAC,IAAI,CAAC,CAAC;QACvB,CAAC;KAAA;CAGF;AAED,MAAM,OAAgB,GAAI,SAAQ,OAAO;IAGjC,SAAS,CAAC,OAAiB,EAAE,IAAe;;YAChD,MAAM,WAAW,GAAG,EAAE,CAAC;YACvB,KAAK,IAAI,CAAC,GAAG,CAAC,EAAE,CAAC,GAAG,OAAO,CAAC,MAAM,EAAE,CAAC,IAAI,CAAC,EAAE;gBAC1C,MAAM,IAAI,GAAG,MAAM,IAAI,CAAC,KAAK,CAAC,OAAO,CAAC,CAAC,CAAC,EAAE,IAAI,CAAC,CAAC;gBAChD,WAAW,CAAC,IAAI,CAAC,CAAC,EAAE,IAAI,EAAE,CAAC,CAAC,CAAC;aAC9B;YACD,OAAO,EAAE,WAAW,EAAE,CAAC;QACzB,CAAC;KAAA;CACF"}
|
|
@@ -0,0 +1,18 @@
|
|
|
1
|
+
export { BaseLLM, LLM, SerializedLLM } from "./base";
|
|
2
|
+
export { OpenAI } from "./openai";
|
|
3
|
+
export { loadLLM } from "./load";
|
|
4
|
+
export type LLMCallbackManager = {
|
|
5
|
+
handleStart: (llm: {
|
|
6
|
+
name: string;
|
|
7
|
+
}, prompts: string[], verbose?: boolean) => void;
|
|
8
|
+
handleError: (err: string, verbose?: boolean) => void;
|
|
9
|
+
handleEnd: (output: LLMResult, verbose?: boolean) => void;
|
|
10
|
+
};
|
|
11
|
+
export type Generation = {
|
|
12
|
+
text: string;
|
|
13
|
+
generationInfo?: Record<string, any>;
|
|
14
|
+
};
|
|
15
|
+
export type LLMResult = {
|
|
16
|
+
generations: Generation[][];
|
|
17
|
+
llmOutput?: Record<string, any>;
|
|
18
|
+
};
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../../llms/index.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,OAAO,EAAE,GAAG,EAAiB,MAAM,QAAQ,CAAC;AACrD,OAAO,EAAE,MAAM,EAAE,MAAM,UAAU,CAAC;AAClC,OAAO,EAAE,OAAO,EAAE,MAAM,QAAQ,CAAC"}
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
{"version":3,"file":"load.js","sourceRoot":"","sources":["../../../llms/load.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,OAAO,EAAE,MAAM,QAAQ,CAAC;AACjC,OAAO,EAAE,eAAe,EAAE,MAAM,SAAS,CAAC;AAE1C,MAAM,CAAC,MAAM,OAAO,GAAG,OAAO,CAAC,WAAW,CAAC;AAC3C,MAAM,CAAC,MAAM,eAAe,GAAG,CAAC,IAAY,EAAE,EAAE,CAAC,OAAO,CAAC,eAAe,CAAC,IAAI,CAAC,CAAC,CAAC"}
|
|
@@ -0,0 +1,63 @@
|
|
|
1
|
+
import type { CreateCompletionRequest } from "openai";
|
|
2
|
+
import { BaseLLM, LLMResult, LLMCallbackManager } from ".";
|
|
3
|
+
interface ModelParams {
|
|
4
|
+
temperature: number;
|
|
5
|
+
maxTokens: number;
|
|
6
|
+
topP: number;
|
|
7
|
+
frequencyPenalty: number;
|
|
8
|
+
presencePenalty: number;
|
|
9
|
+
n: number;
|
|
10
|
+
bestOf: number;
|
|
11
|
+
logitBias?: Record<string, number>;
|
|
12
|
+
}
|
|
13
|
+
type Kwargs = Record<string, any>;
|
|
14
|
+
export declare class OpenAI extends BaseLLM implements ModelParams {
|
|
15
|
+
temperature: number;
|
|
16
|
+
maxTokens: number;
|
|
17
|
+
topP: number;
|
|
18
|
+
frequencyPenalty: number;
|
|
19
|
+
presencePenalty: number;
|
|
20
|
+
n: number;
|
|
21
|
+
bestOf: number;
|
|
22
|
+
logitBias?: Record<string, number>;
|
|
23
|
+
modelName: string;
|
|
24
|
+
modelKwargs?: Kwargs;
|
|
25
|
+
batchSize: number;
|
|
26
|
+
maxRetries: number;
|
|
27
|
+
stop?: string[];
|
|
28
|
+
private client;
|
|
29
|
+
constructor(fields?: Partial<ModelParams> & {
|
|
30
|
+
callbackManager?: LLMCallbackManager;
|
|
31
|
+
verbose?: boolean;
|
|
32
|
+
modelName?: string;
|
|
33
|
+
modelKwargs?: Kwargs;
|
|
34
|
+
openAIApiKey?: string;
|
|
35
|
+
batchSize?: number;
|
|
36
|
+
maxRetries?: number;
|
|
37
|
+
stop?: string[];
|
|
38
|
+
});
|
|
39
|
+
invocationParams(): CreateCompletionRequest & Kwargs;
|
|
40
|
+
identifyingParams(): {
|
|
41
|
+
model: string;
|
|
42
|
+
prompt?: import("openai").CreateCompletionRequestPrompt | null | undefined;
|
|
43
|
+
suffix?: string | null | undefined;
|
|
44
|
+
max_tokens?: number | null | undefined;
|
|
45
|
+
temperature?: number | null | undefined;
|
|
46
|
+
top_p?: number | null | undefined;
|
|
47
|
+
n?: number | null | undefined;
|
|
48
|
+
stream?: boolean | null | undefined;
|
|
49
|
+
logprobs?: number | null | undefined;
|
|
50
|
+
echo?: boolean | null | undefined;
|
|
51
|
+
stop?: import("openai").CreateCompletionRequestStop | null | undefined;
|
|
52
|
+
presence_penalty?: number | null | undefined;
|
|
53
|
+
frequency_penalty?: number | null | undefined;
|
|
54
|
+
best_of?: number | null | undefined;
|
|
55
|
+
logit_bias?: object | null | undefined;
|
|
56
|
+
user?: string | undefined;
|
|
57
|
+
model_name: string;
|
|
58
|
+
};
|
|
59
|
+
_generate(prompts: string[], stop?: string[]): Promise<LLMResult>;
|
|
60
|
+
completionWithRetry(request: CreateCompletionRequest): Promise<import("axios").AxiosResponse<import("openai").CreateCompletionResponse, any>>;
|
|
61
|
+
_llmType(): string;
|
|
62
|
+
}
|
|
63
|
+
export {};
|