langchain 0.0.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (217) hide show
  1. package/dist/cjs/agents/agent.d.ts +42 -0
  2. package/dist/cjs/agents/agent.js +129 -0
  3. package/dist/cjs/agents/agent.js.map +1 -0
  4. package/dist/cjs/agents/executor.d.ts +27 -0
  5. package/dist/cjs/agents/executor.js +105 -0
  6. package/dist/cjs/agents/executor.js.map +1 -0
  7. package/dist/cjs/agents/helpers.d.ts +3 -0
  8. package/dist/cjs/agents/helpers.js +30 -0
  9. package/dist/cjs/agents/helpers.js.map +1 -0
  10. package/dist/cjs/agents/index.d.ts +6 -0
  11. package/dist/cjs/agents/index.js +13 -0
  12. package/dist/cjs/agents/index.js.map +1 -0
  13. package/dist/cjs/agents/load.d.ts +6 -0
  14. package/dist/cjs/agents/load.js +28 -0
  15. package/dist/cjs/agents/load.js.map +1 -0
  16. package/dist/cjs/agents/mrkl/index.d.ts +33 -0
  17. package/dist/cjs/agents/mrkl/index.js +109 -0
  18. package/dist/cjs/agents/mrkl/index.js.map +1 -0
  19. package/dist/cjs/agents/mrkl/prompt.d.ts +3 -0
  20. package/dist/cjs/agents/mrkl/prompt.js +21 -0
  21. package/dist/cjs/agents/mrkl/prompt.js.map +1 -0
  22. package/dist/cjs/agents/tests/agent.test.d.ts +1 -0
  23. package/dist/cjs/agents/tests/agent.test.js +31 -0
  24. package/dist/cjs/agents/tests/agent.test.js.map +1 -0
  25. package/dist/cjs/agents/tools/index.d.ts +7 -0
  26. package/dist/cjs/agents/tools/index.js +6 -0
  27. package/dist/cjs/agents/tools/index.js.map +1 -0
  28. package/dist/cjs/agents/tools/llm-math.d.ts +0 -0
  29. package/dist/cjs/agents/tools/llm-math.js +2 -0
  30. package/dist/cjs/agents/tools/llm-math.js.map +1 -0
  31. package/dist/cjs/agents/tools/serpapi.d.ts +3 -0
  32. package/dist/cjs/agents/tools/serpapi.js +50 -0
  33. package/dist/cjs/agents/tools/serpapi.js.map +1 -0
  34. package/dist/cjs/agents/types.d.ts +25 -0
  35. package/dist/cjs/agents/types.js +3 -0
  36. package/dist/cjs/agents/types.js.map +1 -0
  37. package/dist/cjs/cache.d.ts +11 -0
  38. package/dist/cjs/cache.js +27 -0
  39. package/dist/cjs/cache.js.map +1 -0
  40. package/dist/cjs/chains/base.d.ts +12 -0
  41. package/dist/cjs/chains/base.js +23 -0
  42. package/dist/cjs/chains/base.js.map +1 -0
  43. package/dist/cjs/chains/index.d.ts +3 -0
  44. package/dist/cjs/chains/index.js +10 -0
  45. package/dist/cjs/chains/index.js.map +1 -0
  46. package/dist/cjs/chains/llm_chain.d.ts +30 -0
  47. package/dist/cjs/chains/llm_chain.js +83 -0
  48. package/dist/cjs/chains/llm_chain.js.map +1 -0
  49. package/dist/cjs/chains/llm_math_chain.d.ts +0 -0
  50. package/dist/cjs/chains/llm_math_chain.js +129 -0
  51. package/dist/cjs/chains/llm_math_chain.js.map +1 -0
  52. package/dist/cjs/chains/load.d.ts +2 -0
  53. package/dist/cjs/chains/load.js +29 -0
  54. package/dist/cjs/chains/load.js.map +1 -0
  55. package/dist/cjs/chains/tests/llm_chain.test.d.ts +1 -0
  56. package/dist/cjs/chains/tests/llm_chain.test.js +32 -0
  57. package/dist/cjs/chains/tests/llm_chain.test.js.map +1 -0
  58. package/dist/cjs/index.d.ts +3 -0
  59. package/dist/cjs/index.js +12 -0
  60. package/dist/cjs/index.js.map +1 -0
  61. package/dist/cjs/llms/base.d.ts +23 -0
  62. package/dist/cjs/llms/base.js +159 -0
  63. package/dist/cjs/llms/base.js.map +1 -0
  64. package/dist/cjs/llms/index.d.ts +18 -0
  65. package/dist/cjs/llms/index.js +11 -0
  66. package/dist/cjs/llms/index.js.map +1 -0
  67. package/dist/cjs/llms/load.d.ts +3 -0
  68. package/dist/cjs/llms/load.js +9 -0
  69. package/dist/cjs/llms/load.js.map +1 -0
  70. package/dist/cjs/llms/openai.d.ts +63 -0
  71. package/dist/cjs/llms/openai.js +202 -0
  72. package/dist/cjs/llms/openai.js.map +1 -0
  73. package/dist/cjs/llms/tests/openai.test.d.ts +1 -0
  74. package/dist/cjs/llms/tests/openai.test.js +19 -0
  75. package/dist/cjs/llms/tests/openai.test.js.map +1 -0
  76. package/dist/cjs/prompt/base.d.ts +19 -0
  77. package/dist/cjs/prompt/base.js +54 -0
  78. package/dist/cjs/prompt/base.js.map +1 -0
  79. package/dist/cjs/prompt/few_shot.d.ts +47 -0
  80. package/dist/cjs/prompt/few_shot.js +143 -0
  81. package/dist/cjs/prompt/few_shot.js.map +1 -0
  82. package/dist/cjs/prompt/index.d.ts +3 -0
  83. package/dist/cjs/prompt/index.js +10 -0
  84. package/dist/cjs/prompt/index.js.map +1 -0
  85. package/dist/cjs/prompt/load.d.ts +2 -0
  86. package/dist/cjs/prompt/load.js +25 -0
  87. package/dist/cjs/prompt/load.js.map +1 -0
  88. package/dist/cjs/prompt/parser.d.ts +40 -0
  89. package/dist/cjs/prompt/parser.js +93 -0
  90. package/dist/cjs/prompt/parser.js.map +1 -0
  91. package/dist/cjs/prompt/prompt.d.ts +28 -0
  92. package/dist/cjs/prompt/prompt.js +91 -0
  93. package/dist/cjs/prompt/prompt.js.map +1 -0
  94. package/dist/cjs/prompt/template.d.ts +16 -0
  95. package/dist/cjs/prompt/template.js +87 -0
  96. package/dist/cjs/prompt/template.js.map +1 -0
  97. package/dist/cjs/prompt/tests/load.test.d.ts +1 -0
  98. package/dist/cjs/prompt/tests/load.test.js +30 -0
  99. package/dist/cjs/prompt/tests/load.test.js.map +1 -0
  100. package/dist/cjs/prompt/tests/template.test.d.ts +1 -0
  101. package/dist/cjs/prompt/tests/template.test.js +27 -0
  102. package/dist/cjs/prompt/tests/template.test.js.map +1 -0
  103. package/dist/cjs/util/hub.d.ts +1 -0
  104. package/dist/cjs/util/hub.js +50 -0
  105. package/dist/cjs/util/hub.js.map +1 -0
  106. package/dist/cjs/util/index.d.ts +12 -0
  107. package/dist/cjs/util/index.js +107 -0
  108. package/dist/cjs/util/index.js.map +1 -0
  109. package/dist/esm/agents/agent.d.ts +42 -0
  110. package/dist/esm/agents/agent.js +124 -0
  111. package/dist/esm/agents/agent.js.map +1 -0
  112. package/dist/esm/agents/executor.d.ts +27 -0
  113. package/dist/esm/agents/executor.js +101 -0
  114. package/dist/esm/agents/executor.js.map +1 -0
  115. package/dist/esm/agents/helpers.d.ts +3 -0
  116. package/dist/esm/agents/helpers.js +26 -0
  117. package/dist/esm/agents/helpers.js.map +1 -0
  118. package/dist/esm/agents/index.d.ts +6 -0
  119. package/dist/esm/agents/index.js +5 -0
  120. package/dist/esm/agents/index.js.map +1 -0
  121. package/dist/esm/agents/load.d.ts +6 -0
  122. package/dist/esm/agents/load.js +24 -0
  123. package/dist/esm/agents/load.js.map +1 -0
  124. package/dist/esm/agents/mrkl/index.d.ts +33 -0
  125. package/dist/esm/agents/mrkl/index.js +106 -0
  126. package/dist/esm/agents/mrkl/index.js.map +1 -0
  127. package/dist/esm/agents/mrkl/prompt.d.ts +3 -0
  128. package/dist/esm/agents/mrkl/prompt.js +17 -0
  129. package/dist/esm/agents/mrkl/prompt.js.map +1 -0
  130. package/dist/esm/agents/tests/agent.test.d.ts +1 -0
  131. package/dist/esm/agents/tests/agent.test.js +29 -0
  132. package/dist/esm/agents/tests/agent.test.js.map +1 -0
  133. package/dist/esm/agents/tools/index.d.ts +7 -0
  134. package/dist/esm/agents/tools/index.js +2 -0
  135. package/dist/esm/agents/tools/index.js.map +1 -0
  136. package/dist/esm/agents/tools/llm-math.d.ts +0 -0
  137. package/dist/esm/agents/tools/llm-math.js +2 -0
  138. package/dist/esm/agents/tools/llm-math.js.map +1 -0
  139. package/dist/esm/agents/tools/serpapi.d.ts +3 -0
  140. package/dist/esm/agents/tools/serpapi.js +46 -0
  141. package/dist/esm/agents/tools/serpapi.js.map +1 -0
  142. package/dist/esm/agents/types.d.ts +25 -0
  143. package/dist/esm/agents/types.js +2 -0
  144. package/dist/esm/agents/types.js.map +1 -0
  145. package/dist/esm/cache.d.ts +11 -0
  146. package/dist/esm/cache.js +22 -0
  147. package/dist/esm/cache.js.map +1 -0
  148. package/dist/esm/chains/base.d.ts +12 -0
  149. package/dist/esm/chains/base.js +19 -0
  150. package/dist/esm/chains/base.js.map +1 -0
  151. package/dist/esm/chains/index.d.ts +3 -0
  152. package/dist/esm/chains/index.js +4 -0
  153. package/dist/esm/chains/index.js.map +1 -0
  154. package/dist/esm/chains/llm_chain.d.ts +30 -0
  155. package/dist/esm/chains/llm_chain.js +79 -0
  156. package/dist/esm/chains/llm_chain.js.map +1 -0
  157. package/dist/esm/chains/llm_math_chain.d.ts +0 -0
  158. package/dist/esm/chains/llm_math_chain.js +129 -0
  159. package/dist/esm/chains/llm_math_chain.js.map +1 -0
  160. package/dist/esm/chains/load.d.ts +2 -0
  161. package/dist/esm/chains/load.js +25 -0
  162. package/dist/esm/chains/load.js.map +1 -0
  163. package/dist/esm/chains/tests/llm_chain.test.d.ts +1 -0
  164. package/dist/esm/chains/tests/llm_chain.test.js +30 -0
  165. package/dist/esm/chains/tests/llm_chain.test.js.map +1 -0
  166. package/dist/esm/index.d.ts +3 -0
  167. package/dist/esm/index.js +4 -0
  168. package/dist/esm/index.js.map +1 -0
  169. package/dist/esm/llms/base.d.ts +23 -0
  170. package/dist/esm/llms/base.js +154 -0
  171. package/dist/esm/llms/base.js.map +1 -0
  172. package/dist/esm/llms/index.d.ts +18 -0
  173. package/dist/esm/llms/index.js +4 -0
  174. package/dist/esm/llms/index.js.map +1 -0
  175. package/dist/esm/llms/load.d.ts +3 -0
  176. package/dist/esm/llms/load.js +5 -0
  177. package/dist/esm/llms/load.js.map +1 -0
  178. package/dist/esm/llms/openai.d.ts +63 -0
  179. package/dist/esm/llms/openai.js +198 -0
  180. package/dist/esm/llms/openai.js.map +1 -0
  181. package/dist/esm/llms/tests/openai.test.d.ts +1 -0
  182. package/dist/esm/llms/tests/openai.test.js +17 -0
  183. package/dist/esm/llms/tests/openai.test.js.map +1 -0
  184. package/dist/esm/prompt/base.d.ts +19 -0
  185. package/dist/esm/prompt/base.js +50 -0
  186. package/dist/esm/prompt/base.js.map +1 -0
  187. package/dist/esm/prompt/few_shot.d.ts +47 -0
  188. package/dist/esm/prompt/few_shot.js +139 -0
  189. package/dist/esm/prompt/few_shot.js.map +1 -0
  190. package/dist/esm/prompt/index.d.ts +3 -0
  191. package/dist/esm/prompt/index.js +4 -0
  192. package/dist/esm/prompt/index.js.map +1 -0
  193. package/dist/esm/prompt/load.d.ts +2 -0
  194. package/dist/esm/prompt/load.js +21 -0
  195. package/dist/esm/prompt/load.js.map +1 -0
  196. package/dist/esm/prompt/parser.d.ts +40 -0
  197. package/dist/esm/prompt/parser.js +86 -0
  198. package/dist/esm/prompt/parser.js.map +1 -0
  199. package/dist/esm/prompt/prompt.d.ts +28 -0
  200. package/dist/esm/prompt/prompt.js +87 -0
  201. package/dist/esm/prompt/prompt.js.map +1 -0
  202. package/dist/esm/prompt/template.d.ts +16 -0
  203. package/dist/esm/prompt/template.js +80 -0
  204. package/dist/esm/prompt/template.js.map +1 -0
  205. package/dist/esm/prompt/tests/load.test.d.ts +1 -0
  206. package/dist/esm/prompt/tests/load.test.js +25 -0
  207. package/dist/esm/prompt/tests/load.test.js.map +1 -0
  208. package/dist/esm/prompt/tests/template.test.d.ts +1 -0
  209. package/dist/esm/prompt/tests/template.test.js +25 -0
  210. package/dist/esm/prompt/tests/template.test.js.map +1 -0
  211. package/dist/esm/util/hub.d.ts +1 -0
  212. package/dist/esm/util/hub.js +43 -0
  213. package/dist/esm/util/hub.js.map +1 -0
  214. package/dist/esm/util/index.d.ts +12 -0
  215. package/dist/esm/util/index.js +74 -0
  216. package/dist/esm/util/index.js.map +1 -0
  217. package/package.json +97 -0
@@ -0,0 +1,129 @@
1
+ "use strict";
2
+ /*
3
+ import { BaseChain, ChainValues, LLMChain } from "./index";
4
+
5
+ import { BaseLLM, SerializedLLM } from "../llms";
6
+ import { BasePromptTemplate, SerializedBasePromptTemplate } from "../prompt";
7
+
8
+ import { resolveConfigFromFile } from "../util";
9
+
10
+ export interface LLMMathChainInput {
11
+ prompt: BasePromptTemplate;
12
+ llm: BaseLLM;
13
+ outputKey: string;
14
+ inputKey: string;
15
+ }
16
+
17
+ export type SerializedLLMChain = {
18
+ _type: "llm_chain";
19
+ llm?: SerializedLLM;
20
+ llm_path?: string;
21
+ prompt?: SerializedBasePromptTemplate;
22
+ prompt_path?: string;
23
+ };
24
+
25
+ export class LLMChain extends BaseChain implements LLMMathChainInput {
26
+ prompt: BasePromptTemplate;
27
+
28
+ llm: BaseLLM;
29
+
30
+ inputKey = "answer";
31
+
32
+ outputKey = "question";
33
+
34
+ constructor(fields: {
35
+ prompt: BasePromptTemplate;
36
+ llm: BaseLLM;
37
+ outputKey?: string;
38
+ inputKey?: string;
39
+ }) {
40
+ super();
41
+ this.prompt = fields.prompt;
42
+ this.llm = fields.llm;
43
+ this.outputKey = fields.outputKey ?? this.outputKey;
44
+ this.inputKey = fields.inputKey ?? this.inputKey;
45
+ }
46
+
47
+ async _call(values: ChainValues, stop?: string[]): Promise<ChainValues> {
48
+ const llm = new LLMChain({ prompt: this.prompt, llm: this.llm });
49
+ const p;
50
+ const formattedString = this.prompt.format(values);
51
+ const llmResult = await this.llm.call(formattedString, stop);
52
+ const result = { [this.outputKey]: llmResult };
53
+ return result;
54
+ }
55
+
56
+ async predict(values: ChainValues, stop?: string[]): Promise<string> {
57
+ const output = await this.call(values, stop);
58
+ return output[this.outputKey];
59
+ }
60
+
61
+ _chainType() {
62
+ return "llm_chain" as const;
63
+ }
64
+
65
+ static async deserialize(data: SerializedLLMChain) {
66
+ const serializedLLM = resolveConfigFromFile<"llm", SerializedLLM>(
67
+ "llm",
68
+ data
69
+ );
70
+ const serializedPrompt = resolveConfigFromFile<
71
+ "prompt",
72
+ SerializedBasePromptTemplate
73
+ >("prompt", data);
74
+
75
+ return new LLMChain({
76
+ llm: await BaseLLM.deserialize(serializedLLM),
77
+ prompt: await BasePromptTemplate.deserialize(serializedPrompt),
78
+ });
79
+ }
80
+
81
+ serialize(): SerializedLLMChain {
82
+ return {
83
+ _type: this._chainType(),
84
+ llm: this.llm.serialize(),
85
+ prompt: this.prompt.serialize(),
86
+ };
87
+ }
88
+ }
89
+
90
+ const PROMPT_TEMPLATE = `You are GPT-3, and you can't do math.
91
+ You can do basic math, and your memorization abilities are impressive, but you can't do any complex calculations that a human could not do in their head. You also have an annoying tendency to just make up highly specific, but wrong, answers.
92
+
93
+ So we hooked you up to a Python 3 kernel, and now you can execute code. If anyone gives you a hard math problem, just use this format and we’ll take care of the rest:
94
+
95
+ Question: \${{Question with hard calculation.}}
96
+ \`\`\`python
97
+ \${{Code that prints what you need to know}}
98
+ \`\`\`
99
+ \`\`\`output
100
+ \${{Output of your code}}
101
+ \`\`\`
102
+ Answer: \${{Answer}}
103
+
104
+ Otherwise, use this simpler format:
105
+
106
+ Question: \${{Question without hard calculation}}
107
+ Answer: \${{Answer}}
108
+
109
+ Begin.
110
+
111
+ Question: What is 37593 * 67?
112
+
113
+ \`\`\`python
114
+ print(37593 * 67)
115
+ \`\`\`
116
+ \`\`\`output
117
+ 2518731
118
+ \`\`\`
119
+ Answer: 2518731
120
+
121
+ Question: {question}
122
+ `;
123
+
124
+ PROMPT = PromptTemplate(
125
+ (input_variables = ["question"]),
126
+ (template = _PROMPT_TEMPLATE)
127
+ );
128
+ */
129
+ //# sourceMappingURL=llm_math_chain.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"llm_math_chain.js","sourceRoot":"","sources":["../../../chains/llm_math_chain.ts"],"names":[],"mappings":";AAAA;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EA8HE"}
@@ -0,0 +1,2 @@
1
+ import { BaseChain } from ".";
2
+ export declare const loadChain: (uri: string) => Promise<BaseChain>;
@@ -0,0 +1,25 @@
1
+ var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
2
+ function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
3
+ return new (P || (P = Promise))(function (resolve, reject) {
4
+ function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
5
+ function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
6
+ function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
7
+ step((generator = generator.apply(thisArg, _arguments || [])).next());
8
+ });
9
+ };
10
+ import { BaseChain } from ".";
11
+ import { loadFromHub } from "../util/hub";
12
+ import { parseFileConfig } from "../util";
13
+ const loadChainFromFile = (file) => __awaiter(void 0, void 0, void 0, function* () {
14
+ const serialized = parseFileConfig(file);
15
+ console.log({ serialized });
16
+ return BaseChain.deserialize(serialized);
17
+ });
18
+ export const loadChain = (uri) => __awaiter(void 0, void 0, void 0, function* () {
19
+ const hubResult = yield loadFromHub(uri, loadChainFromFile, "chains", new Set(["json", "yaml"]));
20
+ if (hubResult) {
21
+ return hubResult;
22
+ }
23
+ return loadChainFromFile(uri);
24
+ });
25
+ //# sourceMappingURL=load.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"load.js","sourceRoot":"","sources":["../../../chains/load.ts"],"names":[],"mappings":";;;;;;;;;AAAA,OAAO,EAAE,SAAS,EAAE,MAAM,GAAG,CAAC;AAC9B,OAAO,EAAE,WAAW,EAAE,MAAM,aAAa,CAAC;AAC1C,OAAO,EAAE,eAAe,EAAE,MAAM,SAAS,CAAC;AAE1C,MAAM,iBAAiB,GAAG,CAAO,IAAY,EAAE,EAAE;IAC/C,MAAM,UAAU,GAAG,eAAe,CAAC,IAAI,CAAC,CAAC;IACzC,OAAO,CAAC,GAAG,CAAC,EAAE,UAAU,EAAE,CAAC,CAAC;IAC5B,OAAO,SAAS,CAAC,WAAW,CAAC,UAAU,CAAC,CAAC;AAC3C,CAAC,CAAA,CAAC;AAEF,MAAM,CAAC,MAAM,SAAS,GAAG,CAAO,GAAW,EAAsB,EAAE;IACjE,MAAM,SAAS,GAAG,MAAM,WAAW,CACjC,GAAG,EACH,iBAAiB,EACjB,QAAQ,EACR,IAAI,GAAG,CAAC,CAAC,MAAM,EAAE,MAAM,CAAC,CAAC,CAC1B,CAAC;IACF,IAAI,SAAS,EAAE;QACb,OAAO,SAAS,CAAC;KAClB;IAED,OAAO,iBAAiB,CAAC,GAAG,CAAC,CAAC;AAChC,CAAC,CAAA,CAAC"}
@@ -0,0 +1 @@
1
+ export {};
@@ -0,0 +1,30 @@
1
+ var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
2
+ function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
3
+ return new (P || (P = Promise))(function (resolve, reject) {
4
+ function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
5
+ function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
6
+ function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
7
+ step((generator = generator.apply(thisArg, _arguments || [])).next());
8
+ });
9
+ };
10
+ import { test } from "@jest/globals";
11
+ import { OpenAI } from "../../llms/openai";
12
+ import { PromptTemplate } from "../../prompt";
13
+ import { LLMChain } from "../llm_chain";
14
+ import { loadChain } from "../load";
15
+ test("Test OpenAI", () => __awaiter(void 0, void 0, void 0, function* () {
16
+ const model = new OpenAI({});
17
+ const prompt = new PromptTemplate({
18
+ template: "Print {foo}",
19
+ inputVariables: ["foo"],
20
+ });
21
+ const chain = new LLMChain({ prompt, llm: model });
22
+ const res = yield chain.call({ foo: "my favorite color" });
23
+ console.log({ res });
24
+ }));
25
+ test("Load chain from hub", () => __awaiter(void 0, void 0, void 0, function* () {
26
+ const chain = yield loadChain("lc://chains/hello-world/chain.json");
27
+ const res = yield chain.call({ topic: "my favorite color" });
28
+ console.log({ res });
29
+ }));
30
+ //# sourceMappingURL=llm_chain.test.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"llm_chain.test.js","sourceRoot":"","sources":["../../../../chains/tests/llm_chain.test.ts"],"names":[],"mappings":";;;;;;;;;AAAA,OAAO,EAAE,IAAI,EAAE,MAAM,eAAe,CAAC;AACrC,OAAO,EAAE,MAAM,EAAE,MAAM,mBAAmB,CAAC;AAC3C,OAAO,EAAE,cAAc,EAAE,MAAM,cAAc,CAAC;AAC9C,OAAO,EAAE,QAAQ,EAAE,MAAM,cAAc,CAAC;AACxC,OAAO,EAAE,SAAS,EAAE,MAAM,SAAS,CAAC;AAEpC,IAAI,CAAC,aAAa,EAAE,GAAS,EAAE;IAC7B,MAAM,KAAK,GAAG,IAAI,MAAM,CAAC,EAAE,CAAC,CAAC;IAC7B,MAAM,MAAM,GAAG,IAAI,cAAc,CAAC;QAChC,QAAQ,EAAE,aAAa;QACvB,cAAc,EAAE,CAAC,KAAK,CAAC;KACxB,CAAC,CAAC;IACH,MAAM,KAAK,GAAG,IAAI,QAAQ,CAAC,EAAE,MAAM,EAAE,GAAG,EAAE,KAAK,EAAE,CAAC,CAAC;IACnD,MAAM,GAAG,GAAG,MAAM,KAAK,CAAC,IAAI,CAAC,EAAE,GAAG,EAAE,mBAAmB,EAAE,CAAC,CAAC;IAC3D,OAAO,CAAC,GAAG,CAAC,EAAE,GAAG,EAAE,CAAC,CAAC;AACvB,CAAC,CAAA,CAAC,CAAC;AAEH,IAAI,CAAC,qBAAqB,EAAE,GAAS,EAAE;IACrC,MAAM,KAAK,GAAG,MAAM,SAAS,CAAC,oCAAoC,CAAC,CAAC;IACpE,MAAM,GAAG,GAAG,MAAM,KAAK,CAAC,IAAI,CAAC,EAAE,KAAK,EAAE,mBAAmB,EAAE,CAAC,CAAC;IAC7D,OAAO,CAAC,GAAG,CAAC,EAAE,GAAG,EAAE,CAAC,CAAC;AACvB,CAAC,CAAA,CAAC,CAAC"}
@@ -0,0 +1,3 @@
1
+ export { PromptTemplate, BasePromptTemplate, FewShotPromptTemplate, } from "./prompt";
2
+ export { LLMChain } from "./chains";
3
+ export { OpenAI } from "./llms";
@@ -0,0 +1,4 @@
1
+ export { PromptTemplate, BasePromptTemplate, FewShotPromptTemplate, } from "./prompt";
2
+ export { LLMChain } from "./chains";
3
+ export { OpenAI } from "./llms";
4
+ //# sourceMappingURL=index.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.js","sourceRoot":"","sources":["../../index.ts"],"names":[],"mappings":"AAAA,OAAO,EACL,cAAc,EACd,kBAAkB,EAClB,qBAAqB,GACtB,MAAM,UAAU,CAAC;AAElB,OAAO,EAAE,QAAQ,EAAE,MAAM,UAAU,CAAC;AACpC,OAAO,EAAE,MAAM,EAAE,MAAM,QAAQ,CAAC"}
@@ -0,0 +1,23 @@
1
+ import { LLMCallbackManager, LLMResult } from "./index";
2
+ export type SerializedLLM = {
3
+ _type: string;
4
+ } & Record<string, any>;
5
+ export declare abstract class BaseLLM {
6
+ name: string;
7
+ cache?: boolean;
8
+ callbackManager: LLMCallbackManager;
9
+ verbose?: boolean;
10
+ constructor(callbackManager?: LLMCallbackManager, verbose?: boolean);
11
+ abstract _generate(prompts: string[], stop?: string[]): Promise<LLMResult>;
12
+ _generateUncached(prompts: string[], stop?: string[]): Promise<LLMResult>;
13
+ generate(prompts: string[], stop?: string[]): Promise<LLMResult>;
14
+ call(prompt: string, stop?: string[]): Promise<string>;
15
+ _identifyingParams(): Record<string, any>;
16
+ abstract _llmType(): string;
17
+ serialize(): SerializedLLM;
18
+ static deserialize(data: SerializedLLM): Promise<BaseLLM>;
19
+ }
20
+ export declare abstract class LLM extends BaseLLM {
21
+ abstract _call(prompt: string, stop?: string[]): Promise<string>;
22
+ _generate(prompts: string[], stop?: string[]): Promise<LLMResult>;
23
+ }
@@ -0,0 +1,154 @@
1
+ var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
2
+ function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
3
+ return new (P || (P = Promise))(function (resolve, reject) {
4
+ function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
5
+ function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
6
+ function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
7
+ step((generator = generator.apply(thisArg, _arguments || [])).next());
8
+ });
9
+ };
10
+ var __rest = (this && this.__rest) || function (s, e) {
11
+ var t = {};
12
+ for (var p in s) if (Object.prototype.hasOwnProperty.call(s, p) && e.indexOf(p) < 0)
13
+ t[p] = s[p];
14
+ if (s != null && typeof Object.getOwnPropertySymbols === "function")
15
+ for (var i = 0, p = Object.getOwnPropertySymbols(s); i < p.length; i++) {
16
+ if (e.indexOf(p[i]) < 0 && Object.prototype.propertyIsEnumerable.call(s, p[i]))
17
+ t[p[i]] = s[p[i]];
18
+ }
19
+ return t;
20
+ };
21
+ import { OpenAI } from "./index";
22
+ import { InMemoryCache } from "../cache";
23
+ const getCallbackManager = () => ({
24
+ handleStart: (..._args) => {
25
+ // console.log(args);
26
+ },
27
+ handleEnd: (..._args) => {
28
+ // console.log(args);
29
+ },
30
+ handleError: (..._args) => {
31
+ // console.log(args);
32
+ },
33
+ });
34
+ const getVerbosity = () => true;
35
+ const cache = new InMemoryCache();
36
+ export class BaseLLM {
37
+ constructor(callbackManager, verbose) {
38
+ Object.defineProperty(this, "name", {
39
+ enumerable: true,
40
+ configurable: true,
41
+ writable: true,
42
+ value: void 0
43
+ });
44
+ Object.defineProperty(this, "cache", {
45
+ enumerable: true,
46
+ configurable: true,
47
+ writable: true,
48
+ value: void 0
49
+ });
50
+ Object.defineProperty(this, "callbackManager", {
51
+ enumerable: true,
52
+ configurable: true,
53
+ writable: true,
54
+ value: void 0
55
+ });
56
+ Object.defineProperty(this, "verbose", {
57
+ enumerable: true,
58
+ configurable: true,
59
+ writable: true,
60
+ value: false
61
+ });
62
+ this.callbackManager = callbackManager !== null && callbackManager !== void 0 ? callbackManager : getCallbackManager();
63
+ this.verbose = verbose !== null && verbose !== void 0 ? verbose : getVerbosity();
64
+ }
65
+ _generateUncached(prompts, stop) {
66
+ return __awaiter(this, void 0, void 0, function* () {
67
+ this.callbackManager.handleStart({ name: this.name }, prompts, this.verbose);
68
+ let output;
69
+ try {
70
+ output = yield this._generate(prompts, stop);
71
+ }
72
+ catch (err) {
73
+ this.callbackManager.handleError(`${err}`, this.verbose);
74
+ throw err;
75
+ }
76
+ this.callbackManager.handleEnd(output, this.verbose);
77
+ return output;
78
+ });
79
+ }
80
+ generate(prompts, stop) {
81
+ var _a;
82
+ return __awaiter(this, void 0, void 0, function* () {
83
+ if (!Array.isArray(prompts)) {
84
+ throw new Error("Argument 'prompts' is expected to be a string[]");
85
+ }
86
+ if (this.cache === true && cache === null) {
87
+ throw new Error("Requested cache, but no cache found");
88
+ }
89
+ if (cache === null || this.cache === false) {
90
+ return this._generateUncached(prompts, stop);
91
+ }
92
+ const params = this.serialize();
93
+ params.stop = stop;
94
+ const llmStringKey = `${Object.entries(params).sort()}`;
95
+ const missingPromptIndices = [];
96
+ const generations = prompts.map((prompt, index) => {
97
+ const result = cache.lookup(prompt, llmStringKey);
98
+ if (!result) {
99
+ missingPromptIndices.push(index);
100
+ }
101
+ return result;
102
+ });
103
+ let llmOutput = {};
104
+ if (missingPromptIndices.length > 0) {
105
+ const results = yield this._generateUncached(missingPromptIndices.map((i) => prompts[i]), stop);
106
+ results.generations.forEach((generation, index) => {
107
+ const promptIndex = missingPromptIndices[index];
108
+ generations[promptIndex] = generation;
109
+ cache.update(prompts[promptIndex], llmStringKey, generation);
110
+ });
111
+ llmOutput = (_a = results.llmOutput) !== null && _a !== void 0 ? _a : {};
112
+ }
113
+ return { generations, llmOutput };
114
+ });
115
+ }
116
+ call(prompt, stop) {
117
+ return __awaiter(this, void 0, void 0, function* () {
118
+ const { generations } = yield this.generate([prompt], stop);
119
+ return generations[0][0].text;
120
+ });
121
+ }
122
+ // eslint-disable-next-line @typescript-eslint/no-explicit-any
123
+ _identifyingParams() {
124
+ return {};
125
+ }
126
+ serialize() {
127
+ return Object.assign(Object.assign({}, this._identifyingParams()), { _type: this._llmType() });
128
+ }
129
+ static deserialize(data) {
130
+ return __awaiter(this, void 0, void 0, function* () {
131
+ const { _type } = data, rest = __rest(data, ["_type"]);
132
+ const Cls = {
133
+ openai: OpenAI,
134
+ }[_type];
135
+ if (Cls === undefined) {
136
+ throw new Error(`Cannot load LLM with type ${_type}`);
137
+ }
138
+ return new Cls(rest);
139
+ });
140
+ }
141
+ }
142
+ export class LLM extends BaseLLM {
143
+ _generate(prompts, stop) {
144
+ return __awaiter(this, void 0, void 0, function* () {
145
+ const generations = [];
146
+ for (let i = 0; i < prompts.length; i += 1) {
147
+ const text = yield this._call(prompts[i], stop);
148
+ generations.push([{ text }]);
149
+ }
150
+ return { generations };
151
+ });
152
+ }
153
+ }
154
+ //# sourceMappingURL=base.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"base.js","sourceRoot":"","sources":["../../../llms/base.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA,OAAO,EAAiC,MAAM,EAAE,MAAM,SAAS,CAAC;AAChE,OAAO,EAAa,aAAa,EAAE,MAAM,UAAU,CAAC;AAEpD,MAAM,kBAAkB,GAAG,GAAuB,EAAE,CAAC,CAAC;IACpD,WAAW,EAAE,CAAC,GAAG,KAAK,EAAE,EAAE;QACxB,qBAAqB;IACvB,CAAC;IACD,SAAS,EAAE,CAAC,GAAG,KAAK,EAAE,EAAE;QACtB,qBAAqB;IACvB,CAAC;IACD,WAAW,EAAE,CAAC,GAAG,KAAK,EAAE,EAAE;QACxB,qBAAqB;IACvB,CAAC;CACF,CAAC,CAAC;AAEH,MAAM,YAAY,GAAG,GAAG,EAAE,CAAC,IAAI,CAAC;AAEhC,MAAM,KAAK,GAAc,IAAI,aAAa,EAAE,CAAC;AAO7C,MAAM,OAAgB,OAAO;IAS3B,YAAY,eAAoC,EAAE,OAAiB;QARnE;;;;;WAAa;QAEb;;;;;WAAgB;QAEhB;;;;;WAAoC;QAEpC;;;;mBAAoB,KAAK;WAAC;QAGxB,IAAI,CAAC,eAAe,GAAG,eAAe,aAAf,eAAe,cAAf,eAAe,GAAI,kBAAkB,EAAE,CAAC;QAC/D,IAAI,CAAC,OAAO,GAAG,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,YAAY,EAAE,CAAC;IAC3C,CAAC;IAIK,iBAAiB,CACrB,OAAiB,EACjB,IAAe;;YAEf,IAAI,CAAC,eAAe,CAAC,WAAW,CAC9B,EAAE,IAAI,EAAE,IAAI,CAAC,IAAI,EAAE,EACnB,OAAO,EACP,IAAI,CAAC,OAAO,CACb,CAAC;YACF,IAAI,MAAM,CAAC;YACX,IAAI;gBACF,MAAM,GAAG,MAAM,IAAI,CAAC,SAAS,CAAC,OAAO,EAAE,IAAI,CAAC,CAAC;aAC9C;YAAC,OAAO,GAAG,EAAE;gBACZ,IAAI,CAAC,eAAe,CAAC,WAAW,CAAC,GAAG,GAAG,EAAE,EAAE,IAAI,CAAC,OAAO,CAAC,CAAC;gBACzD,MAAM,GAAG,CAAC;aACX;YAED,IAAI,CAAC,eAAe,CAAC,SAAS,CAAC,MAAM,EAAE,IAAI,CAAC,OAAO,CAAC,CAAC;YACrD,OAAO,MAAM,CAAC;QAChB,CAAC;KAAA;IAEK,QAAQ,CAAC,OAAiB,EAAE,IAAe;;;YAC/C,IAAI,CAAC,KAAK,CAAC,OAAO,CAAC,OAAO,CAAC,EAAE;gBAC3B,MAAM,IAAI,KAAK,CAAC,iDAAiD,CAAC,CAAC;aACpE;YAED,IAAI,IAAI,CAAC,KAAK,KAAK,IAAI,IAAI,KAAK,KAAK,IAAI,EAAE;gBACzC,MAAM,IAAI,KAAK,CAAC,qCAAqC,CAAC,CAAC;aACxD;YAED,IAAI,KAAK,KAAK,IAAI,IAAI,IAAI,CAAC,KAAK,KAAK,KAAK,EAAE;gBAC1C,OAAO,IAAI,CAAC,iBAAiB,CAAC,OAAO,EAAE,IAAI,CAAC,CAAC;aAC9C;YAED,MAAM,MAAM,GAAG,IAAI,CAAC,SAAS,EAAE,CAAC;YAChC,MAAM,CAAC,IAAI,GAAG,IAAI,CAAC;YAEnB,MAAM,YAAY,GAAG,GAAG,MAAM,CAAC,OAAO,CAAC,MAAM,CAAC,CAAC,IAAI,EAAE,EAAE,CAAC;YACxD,MAAM,oBAAoB,GAAa,EAAE,CAAC;YAC1C,MAAM,WAAW,GAAG,OAAO,CAAC,GAAG,CAAC,CAAC,MAAM,EAAE,KAAK,EAAE,EAAE;gBAChD,MAAM,MAAM,GAAG,KAAK,CAAC,MAAM,CAAC,MAAM,EAAE,YAAY,CAAC,CAAC;gBAClD,IAAI,CAAC,MAAM,EAAE;oBACX,oBAAoB,CAAC,IAAI,CAAC,KAAK,CAAC,CAAC;iBAClC;gBACD,OAAO,MAAM,CAAC;YAChB,CAAC,CAAC,CAAC;YAEH,IAAI,SAAS,GAAG,EAAE,CAAC;YACnB,IAAI,oBAAoB,CAAC,MAAM,GAAG,CAAC,EAAE;gBACnC,MAAM,OAAO,GAAG,MAAM,IAAI,CAAC,iBAAiB,CAC1C,oBAAoB,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE,EAAE,CAAC,OAAO,CAAC,CAAC,CAAC,CAAC,EAC3C,IAAI,CACL,CAAC;gBACF,OAAO,CAAC,WAAW,CAAC,OAAO,CAAC,CAAC,UAAU,EAAE,KAAK,EAAE,EAAE;oBAChD,MAAM,WAAW,GAAG,oBAAoB,CAAC,KAAK,CAAC,CAAC;oBAChD,WAAW,CAAC,WAAW,CAAC,GAAG,UAAU,CAAC;oBACtC,KAAK,CAAC,MAAM,CAAC,OAAO,CAAC,WAAW,CAAC,EAAE,YAAY,EAAE,UAAU,CAAC,CAAC;gBAC/D,CAAC,CAAC,CAAC;gBACH,SAAS,GAAG,MAAA,OAAO,CAAC,SAAS,mCAAI,EAAE,CAAC;aACrC;YAED,OAAO,EAAE,WAAW,EAAE,SAAS,EAAe,CAAC;;KAChD;IAEK,IAAI,CAAC,MAAc,EAAE,IAAe;;YACxC,MAAM,EAAE,WAAW,EAAE,GAAG,MAAM,IAAI,CAAC,QAAQ,CAAC,CAAC,MAAM,CAAC,EAAE,IAAI,CAAC,CAAC;YAC5D,OAAO,WAAW,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,IAAI,CAAC;QAChC,CAAC;KAAA;IAED,8DAA8D;IAC9D,kBAAkB;QAChB,OAAO,EAAE,CAAC;IACZ,CAAC;IAID,SAAS;QACP,uCACK,IAAI,CAAC,kBAAkB,EAAE,KAC5B,KAAK,EAAE,IAAI,CAAC,QAAQ,EAAE,IACtB;IACJ,CAAC;IAED,MAAM,CAAO,WAAW,CAAC,IAAmB;;YAC1C,MAAM,EAAE,KAAK,KAAc,IAAI,EAAb,IAAI,UAAK,IAAI,EAAzB,SAAkB,CAAO,CAAC;YAChC,MAAM,GAAG,GAAG;gBACV,MAAM,EAAE,MAAM;aACf,CAAC,KAAK,CAAC,CAAC;YACT,IAAI,GAAG,KAAK,SAAS,EAAE;gBACrB,MAAM,IAAI,KAAK,CAAC,8BAA8B,KAAK,EAAE,CAAC,CAAC;aACxD;YACD,OAAO,IAAI,GAAG,CAAC,IAAI,CAAC,CAAC;QACvB,CAAC;KAAA;CAGF;AAED,MAAM,OAAgB,GAAI,SAAQ,OAAO;IAGjC,SAAS,CAAC,OAAiB,EAAE,IAAe;;YAChD,MAAM,WAAW,GAAG,EAAE,CAAC;YACvB,KAAK,IAAI,CAAC,GAAG,CAAC,EAAE,CAAC,GAAG,OAAO,CAAC,MAAM,EAAE,CAAC,IAAI,CAAC,EAAE;gBAC1C,MAAM,IAAI,GAAG,MAAM,IAAI,CAAC,KAAK,CAAC,OAAO,CAAC,CAAC,CAAC,EAAE,IAAI,CAAC,CAAC;gBAChD,WAAW,CAAC,IAAI,CAAC,CAAC,EAAE,IAAI,EAAE,CAAC,CAAC,CAAC;aAC9B;YACD,OAAO,EAAE,WAAW,EAAE,CAAC;QACzB,CAAC;KAAA;CACF"}
@@ -0,0 +1,18 @@
1
+ export { BaseLLM, LLM, SerializedLLM } from "./base";
2
+ export { OpenAI } from "./openai";
3
+ export { loadLLM } from "./load";
4
+ export type LLMCallbackManager = {
5
+ handleStart: (llm: {
6
+ name: string;
7
+ }, prompts: string[], verbose?: boolean) => void;
8
+ handleError: (err: string, verbose?: boolean) => void;
9
+ handleEnd: (output: LLMResult, verbose?: boolean) => void;
10
+ };
11
+ export type Generation = {
12
+ text: string;
13
+ generationInfo?: Record<string, any>;
14
+ };
15
+ export type LLMResult = {
16
+ generations: Generation[][];
17
+ llmOutput?: Record<string, any>;
18
+ };
@@ -0,0 +1,4 @@
1
+ export { BaseLLM, LLM } from "./base";
2
+ export { OpenAI } from "./openai";
3
+ export { loadLLM } from "./load";
4
+ //# sourceMappingURL=index.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.js","sourceRoot":"","sources":["../../../llms/index.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,OAAO,EAAE,GAAG,EAAiB,MAAM,QAAQ,CAAC;AACrD,OAAO,EAAE,MAAM,EAAE,MAAM,UAAU,CAAC;AAClC,OAAO,EAAE,OAAO,EAAE,MAAM,QAAQ,CAAC"}
@@ -0,0 +1,3 @@
1
+ import { BaseLLM } from "./base";
2
+ export declare const loadLLM: typeof BaseLLM.deserialize;
3
+ export declare const loadLLMFromFile: (file: string) => Promise<BaseLLM>;
@@ -0,0 +1,5 @@
1
+ import { BaseLLM } from "./base";
2
+ import { parseFileConfig } from "../util";
3
+ export const loadLLM = BaseLLM.deserialize;
4
+ export const loadLLMFromFile = (file) => loadLLM(parseFileConfig(file));
5
+ //# sourceMappingURL=load.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"load.js","sourceRoot":"","sources":["../../../llms/load.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,OAAO,EAAE,MAAM,QAAQ,CAAC;AACjC,OAAO,EAAE,eAAe,EAAE,MAAM,SAAS,CAAC;AAE1C,MAAM,CAAC,MAAM,OAAO,GAAG,OAAO,CAAC,WAAW,CAAC;AAC3C,MAAM,CAAC,MAAM,eAAe,GAAG,CAAC,IAAY,EAAE,EAAE,CAAC,OAAO,CAAC,eAAe,CAAC,IAAI,CAAC,CAAC,CAAC"}
@@ -0,0 +1,63 @@
1
+ import type { CreateCompletionRequest } from "openai";
2
+ import { BaseLLM, LLMResult, LLMCallbackManager } from ".";
3
+ interface ModelParams {
4
+ temperature: number;
5
+ maxTokens: number;
6
+ topP: number;
7
+ frequencyPenalty: number;
8
+ presencePenalty: number;
9
+ n: number;
10
+ bestOf: number;
11
+ logitBias?: Record<string, number>;
12
+ }
13
+ type Kwargs = Record<string, any>;
14
+ export declare class OpenAI extends BaseLLM implements ModelParams {
15
+ temperature: number;
16
+ maxTokens: number;
17
+ topP: number;
18
+ frequencyPenalty: number;
19
+ presencePenalty: number;
20
+ n: number;
21
+ bestOf: number;
22
+ logitBias?: Record<string, number>;
23
+ modelName: string;
24
+ modelKwargs?: Kwargs;
25
+ batchSize: number;
26
+ maxRetries: number;
27
+ stop?: string[];
28
+ private client;
29
+ constructor(fields?: Partial<ModelParams> & {
30
+ callbackManager?: LLMCallbackManager;
31
+ verbose?: boolean;
32
+ modelName?: string;
33
+ modelKwargs?: Kwargs;
34
+ openAIApiKey?: string;
35
+ batchSize?: number;
36
+ maxRetries?: number;
37
+ stop?: string[];
38
+ });
39
+ invocationParams(): CreateCompletionRequest & Kwargs;
40
+ identifyingParams(): {
41
+ model: string;
42
+ prompt?: import("openai").CreateCompletionRequestPrompt | null | undefined;
43
+ suffix?: string | null | undefined;
44
+ max_tokens?: number | null | undefined;
45
+ temperature?: number | null | undefined;
46
+ top_p?: number | null | undefined;
47
+ n?: number | null | undefined;
48
+ stream?: boolean | null | undefined;
49
+ logprobs?: number | null | undefined;
50
+ echo?: boolean | null | undefined;
51
+ stop?: import("openai").CreateCompletionRequestStop | null | undefined;
52
+ presence_penalty?: number | null | undefined;
53
+ frequency_penalty?: number | null | undefined;
54
+ best_of?: number | null | undefined;
55
+ logit_bias?: object | null | undefined;
56
+ user?: string | undefined;
57
+ model_name: string;
58
+ };
59
+ _generate(prompts: string[], stop?: string[]): Promise<LLMResult>;
60
+ completionWithRetry(request: CreateCompletionRequest): Promise<import("axios").AxiosResponse<import("openai").CreateCompletionResponse, any>>;
61
+ _llmType(): string;
62
+ }
63
+ export {};