@botpress/zai 1.0.0-beta.2 → 1.0.0-beta.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +1898 -0
- package/dist/index.cjs.map +1 -0
- package/dist/index.d.cts +916 -0
- package/dist/index.d.ts +916 -0
- package/dist/index.js +1868 -0
- package/dist/index.js.map +1 -0
- package/ensure-env.cjs +9 -0
- package/package.json +25 -17
- package/tsconfig.json +23 -18
- package/tsup.config.ts +16 -0
- package/vitest.config.ts +2 -16
- package/vitest.setup.ts +11 -15
- package/dist/src/adapters/adapter.d.ts +0 -27
- package/dist/src/adapters/adapter.js +0 -3
- package/dist/src/adapters/adapter.js.map +0 -1
- package/dist/src/adapters/botpress-table.d.ts +0 -155
- package/dist/src/adapters/botpress-table.js +0 -170
- package/dist/src/adapters/botpress-table.js.map +0 -1
- package/dist/src/adapters/memory.d.ts +0 -7
- package/dist/src/adapters/memory.js +0 -13
- package/dist/src/adapters/memory.js.map +0 -1
- package/dist/src/env.d.ts +0 -1
- package/dist/src/env.js +0 -27
- package/dist/src/env.js.map +0 -1
- package/dist/src/index.d.ts +0 -9
- package/dist/src/index.js +0 -10
- package/dist/src/index.js.map +0 -1
- package/dist/src/models.d.ts +0 -253
- package/dist/src/models.js +0 -348
- package/dist/src/models.js.map +0 -1
- package/dist/src/operations/__tests/index.d.ts +0 -19
- package/dist/src/operations/__tests/index.js +0 -24
- package/dist/src/operations/__tests/index.js.map +0 -1
- package/dist/src/operations/check.d.ts +0 -36
- package/dist/src/operations/check.js +0 -146
- package/dist/src/operations/check.js.map +0 -1
- package/dist/src/operations/check.test.d.ts +0 -1
- package/dist/src/operations/check.test.js +0 -129
- package/dist/src/operations/check.test.js.map +0 -1
- package/dist/src/operations/constants.d.ts +0 -2
- package/dist/src/operations/constants.js +0 -3
- package/dist/src/operations/constants.js.map +0 -1
- package/dist/src/operations/errors.d.ts +0 -5
- package/dist/src/operations/errors.js +0 -11
- package/dist/src/operations/errors.js.map +0 -1
- package/dist/src/operations/extract.d.ts +0 -20
- package/dist/src/operations/extract.js +0 -235
- package/dist/src/operations/extract.js.map +0 -1
- package/dist/src/operations/extract.test.d.ts +0 -1
- package/dist/src/operations/extract.test.js +0 -163
- package/dist/src/operations/extract.test.js.map +0 -1
- package/dist/src/operations/filter.d.ts +0 -39
- package/dist/src/operations/filter.js +0 -184
- package/dist/src/operations/filter.js.map +0 -1
- package/dist/src/operations/filter.test.d.ts +0 -1
- package/dist/src/operations/filter.test.js +0 -151
- package/dist/src/operations/filter.test.js.map +0 -1
- package/dist/src/operations/label.d.ts +0 -79
- package/dist/src/operations/label.js +0 -250
- package/dist/src/operations/label.js.map +0 -1
- package/dist/src/operations/label.test.d.ts +0 -1
- package/dist/src/operations/label.test.js +0 -213
- package/dist/src/operations/label.test.js.map +0 -1
- package/dist/src/operations/rewrite.d.ts +0 -34
- package/dist/src/operations/rewrite.js +0 -106
- package/dist/src/operations/rewrite.js.map +0 -1
- package/dist/src/operations/rewrite.test.d.ts +0 -1
- package/dist/src/operations/rewrite.test.js +0 -96
- package/dist/src/operations/rewrite.test.js.map +0 -1
- package/dist/src/operations/summarize.d.ts +0 -46
- package/dist/src/operations/summarize.js +0 -139
- package/dist/src/operations/summarize.js.map +0 -1
- package/dist/src/operations/summarize.test.d.ts +0 -1
- package/dist/src/operations/summarize.test.js +0 -22
- package/dist/src/operations/summarize.test.js.map +0 -1
- package/dist/src/operations/text.d.ts +0 -16
- package/dist/src/operations/text.js +0 -47
- package/dist/src/operations/text.js.map +0 -1
- package/dist/src/operations/text.test.d.ts +0 -1
- package/dist/src/operations/text.test.js +0 -51
- package/dist/src/operations/text.test.js.map +0 -1
- package/dist/src/operations/zai-learn.test.d.ts +0 -1
- package/dist/src/operations/zai-learn.test.js +0 -71
- package/dist/src/operations/zai-learn.test.js.map +0 -1
- package/dist/src/operations/zai-retry.test.d.ts +0 -1
- package/dist/src/operations/zai-retry.test.js +0 -50
- package/dist/src/operations/zai-retry.test.js.map +0 -1
- package/dist/src/utils.d.ts +0 -53
- package/dist/src/utils.js +0 -49
- package/dist/src/utils.js.map +0 -1
- package/dist/src/zai.d.ts +0 -91
- package/dist/src/zai.js +0 -139
- package/dist/src/zai.js.map +0 -1
- package/dist/tsconfig.tsbuildinfo +0 -1
- package/dist/typings.d.ts +0 -901
- package/dist/typings.tsbuildinfo +0 -1
- package/dist/vitest.config.d.ts +0 -2
- package/dist/vitest.config.js +0 -21
- package/dist/vitest.config.js.map +0 -1
- package/dist/vitest.setup.d.ts +0 -1
- package/dist/vitest.setup.js +0 -22
- package/dist/vitest.setup.js.map +0 -1
- package/src/adapters/adapter.ts +0 -35
- package/src/adapters/botpress-table.ts +0 -213
- package/src/adapters/memory.ts +0 -13
- package/src/env.ts +0 -54
- package/src/index.ts +0 -11
- package/src/models.ts +0 -347
- package/src/operations/__tests/botpress_docs.txt +0 -26040
- package/src/operations/__tests/index.ts +0 -30
- package/src/operations/check.test.ts +0 -155
- package/src/operations/check.ts +0 -187
- package/src/operations/constants.ts +0 -2
- package/src/operations/errors.ts +0 -9
- package/src/operations/extract.test.ts +0 -209
- package/src/operations/extract.ts +0 -291
- package/src/operations/filter.test.ts +0 -182
- package/src/operations/filter.ts +0 -231
- package/src/operations/label.test.ts +0 -239
- package/src/operations/label.ts +0 -332
- package/src/operations/rewrite.test.ts +0 -114
- package/src/operations/rewrite.ts +0 -148
- package/src/operations/summarize.test.ts +0 -25
- package/src/operations/summarize.ts +0 -193
- package/src/operations/text.test.ts +0 -60
- package/src/operations/text.ts +0 -63
- package/src/operations/zai-learn.test.ts +0 -85
- package/src/operations/zai-retry.test.ts +0 -64
- package/src/scripts/update-models.ts +0 -74
- package/src/utils.ts +0 -61
- package/src/zai.ts +0 -185
- package/tsconfig.types.json +0 -22
- package/turbo.json +0 -14
package/dist/src/models.d.ts
DELETED
|
@@ -1,253 +0,0 @@
|
|
|
1
|
-
export declare const Models: readonly [{
|
|
2
|
-
readonly id: "anthropic__claude-3-haiku-20240307";
|
|
3
|
-
readonly name: "Claude 3 Haiku";
|
|
4
|
-
readonly description: "Claude 3 Haiku is Anthropic's fastest and most compact model for near-instant responsiveness. Quick and accurate targeted performance.";
|
|
5
|
-
readonly tags: readonly ["low-cost", "general-purpose"];
|
|
6
|
-
readonly input: {
|
|
7
|
-
readonly costPer1MTokens: 0.25;
|
|
8
|
-
readonly maxTokens: 200000;
|
|
9
|
-
};
|
|
10
|
-
readonly output: {
|
|
11
|
-
readonly costPer1MTokens: 1.25;
|
|
12
|
-
readonly maxTokens: 4096;
|
|
13
|
-
};
|
|
14
|
-
readonly integration: "anthropic";
|
|
15
|
-
}, {
|
|
16
|
-
readonly id: "anthropic__claude-3-5-sonnet-20240620";
|
|
17
|
-
readonly name: "Claude 3.5 Sonnet";
|
|
18
|
-
readonly description: "Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at coding, data science, visual processing, and agentic tasks.";
|
|
19
|
-
readonly tags: readonly ["recommended", "vision", "general-purpose", "agents", "coding", "function-calling", "storytelling"];
|
|
20
|
-
readonly input: {
|
|
21
|
-
readonly costPer1MTokens: 3;
|
|
22
|
-
readonly maxTokens: 200000;
|
|
23
|
-
};
|
|
24
|
-
readonly output: {
|
|
25
|
-
readonly costPer1MTokens: 15;
|
|
26
|
-
readonly maxTokens: 4096;
|
|
27
|
-
};
|
|
28
|
-
readonly integration: "anthropic";
|
|
29
|
-
}, {
|
|
30
|
-
readonly id: "groq__gemma-7b-it";
|
|
31
|
-
readonly name: "Gemma 7B";
|
|
32
|
-
readonly description: "Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning.";
|
|
33
|
-
readonly tags: readonly ["deprecated", "low-cost"];
|
|
34
|
-
readonly input: {
|
|
35
|
-
readonly costPer1MTokens: 0.07;
|
|
36
|
-
readonly maxTokens: 8192;
|
|
37
|
-
};
|
|
38
|
-
readonly output: {
|
|
39
|
-
readonly costPer1MTokens: 0.07;
|
|
40
|
-
readonly maxTokens: 8192;
|
|
41
|
-
};
|
|
42
|
-
readonly integration: "groq";
|
|
43
|
-
}, {
|
|
44
|
-
readonly id: "groq__gemma2-9b-it";
|
|
45
|
-
readonly name: "Gemma2 9B";
|
|
46
|
-
readonly description: "Redesigned for outsized performance and unmatched efficiency, Gemma 2 optimizes for blazing-fast inference on diverse hardware. Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning.";
|
|
47
|
-
readonly tags: readonly ["low-cost", "general-purpose"];
|
|
48
|
-
readonly input: {
|
|
49
|
-
readonly costPer1MTokens: 0.2;
|
|
50
|
-
readonly maxTokens: 8192;
|
|
51
|
-
};
|
|
52
|
-
readonly output: {
|
|
53
|
-
readonly costPer1MTokens: 0.2;
|
|
54
|
-
readonly maxTokens: 8192;
|
|
55
|
-
};
|
|
56
|
-
readonly integration: "groq";
|
|
57
|
-
}, {
|
|
58
|
-
readonly id: "groq__llama3-70b-8192";
|
|
59
|
-
readonly name: "LLaMA 3 70B";
|
|
60
|
-
readonly tags: readonly ["general-purpose"];
|
|
61
|
-
readonly description: "Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. The Llama 3 instruction tuned models are optimized for dialogue use cases and outperform many of the available open source chat models on common industry benchmarks.";
|
|
62
|
-
readonly input: {
|
|
63
|
-
readonly costPer1MTokens: 0.59;
|
|
64
|
-
readonly maxTokens: 8192;
|
|
65
|
-
};
|
|
66
|
-
readonly output: {
|
|
67
|
-
readonly costPer1MTokens: 0.79;
|
|
68
|
-
readonly maxTokens: 8192;
|
|
69
|
-
};
|
|
70
|
-
readonly integration: "groq";
|
|
71
|
-
}, {
|
|
72
|
-
readonly id: "groq__llama3-8b-8192";
|
|
73
|
-
readonly name: "LLaMA 3 8B";
|
|
74
|
-
readonly description: "Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. The Llama 3 instruction tuned models are optimized for dialogue use cases and outperform many of the available open source chat models on common industry benchmarks.";
|
|
75
|
-
readonly tags: readonly ["low-cost", "general-purpose", "deprecated"];
|
|
76
|
-
readonly input: {
|
|
77
|
-
readonly costPer1MTokens: 0.05;
|
|
78
|
-
readonly maxTokens: 8192;
|
|
79
|
-
};
|
|
80
|
-
readonly output: {
|
|
81
|
-
readonly costPer1MTokens: 0.08;
|
|
82
|
-
readonly maxTokens: 8192;
|
|
83
|
-
};
|
|
84
|
-
readonly integration: "groq";
|
|
85
|
-
}, {
|
|
86
|
-
readonly id: "groq__llama-3.1-70b-versatile";
|
|
87
|
-
readonly name: "LLaMA 3.1 70B";
|
|
88
|
-
readonly description: "The Llama 3.1 instruction-tuned, text-only models are optimized for multilingual dialogue use cases.";
|
|
89
|
-
readonly tags: readonly ["recommended", "general-purpose"];
|
|
90
|
-
readonly input: {
|
|
91
|
-
readonly costPer1MTokens: 0.59;
|
|
92
|
-
readonly maxTokens: 128000;
|
|
93
|
-
};
|
|
94
|
-
readonly output: {
|
|
95
|
-
readonly costPer1MTokens: 0.79;
|
|
96
|
-
readonly maxTokens: 8192;
|
|
97
|
-
};
|
|
98
|
-
readonly integration: "groq";
|
|
99
|
-
}, {
|
|
100
|
-
readonly id: "groq__llama-3.1-8b-instant";
|
|
101
|
-
readonly name: "LLaMA 3.1 8B";
|
|
102
|
-
readonly description: "The Llama 3.1 instruction-tuned, text-only models are optimized for multilingual dialogue use cases.";
|
|
103
|
-
readonly tags: readonly ["low-cost", "general-purpose"];
|
|
104
|
-
readonly input: {
|
|
105
|
-
readonly costPer1MTokens: 0.05;
|
|
106
|
-
readonly maxTokens: 128000;
|
|
107
|
-
};
|
|
108
|
-
readonly output: {
|
|
109
|
-
readonly costPer1MTokens: 0.08;
|
|
110
|
-
readonly maxTokens: 8192;
|
|
111
|
-
};
|
|
112
|
-
readonly integration: "groq";
|
|
113
|
-
}, {
|
|
114
|
-
readonly id: "groq__llama-3.2-11b-vision-preview";
|
|
115
|
-
readonly name: "LLaMA 3.2 11B Vision (Preview)";
|
|
116
|
-
readonly description: "[BETA preview version] The Llama 3.2-Vision instruction-tuned models are optimized for visual recognition, image reasoning, captioning, and answering general questions about an image.";
|
|
117
|
-
readonly tags: readonly ["low-cost", "vision", "general-purpose"];
|
|
118
|
-
readonly input: {
|
|
119
|
-
readonly costPer1MTokens: 0.18;
|
|
120
|
-
readonly maxTokens: 128000;
|
|
121
|
-
};
|
|
122
|
-
readonly output: {
|
|
123
|
-
readonly costPer1MTokens: 0.18;
|
|
124
|
-
readonly maxTokens: 8192;
|
|
125
|
-
};
|
|
126
|
-
readonly integration: "groq";
|
|
127
|
-
}, {
|
|
128
|
-
readonly id: "groq__llama-3.2-1b-preview";
|
|
129
|
-
readonly name: "LLaMA 3.2 1B (Preview)";
|
|
130
|
-
readonly description: "[BETA preview version] The Llama 3.2 instruction-tuned, text-only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
|
|
131
|
-
readonly tags: readonly ["low-cost"];
|
|
132
|
-
readonly input: {
|
|
133
|
-
readonly costPer1MTokens: 0.04;
|
|
134
|
-
readonly maxTokens: 128000;
|
|
135
|
-
};
|
|
136
|
-
readonly output: {
|
|
137
|
-
readonly costPer1MTokens: 0.04;
|
|
138
|
-
readonly maxTokens: 8192;
|
|
139
|
-
};
|
|
140
|
-
readonly integration: "groq";
|
|
141
|
-
}, {
|
|
142
|
-
readonly id: "groq__llama-3.2-3b-preview";
|
|
143
|
-
readonly name: "LLaMA 3.2 3B (Preview)";
|
|
144
|
-
readonly description: "[BETA preview version] The Llama 3.2 instruction-tuned, text-only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.";
|
|
145
|
-
readonly tags: readonly ["low-cost", "general-purpose"];
|
|
146
|
-
readonly input: {
|
|
147
|
-
readonly costPer1MTokens: 0.06;
|
|
148
|
-
readonly maxTokens: 128000;
|
|
149
|
-
};
|
|
150
|
-
readonly output: {
|
|
151
|
-
readonly costPer1MTokens: 0.06;
|
|
152
|
-
readonly maxTokens: 8192;
|
|
153
|
-
};
|
|
154
|
-
readonly integration: "groq";
|
|
155
|
-
}, {
|
|
156
|
-
readonly id: "groq__llama-3.2-90b-vision-preview";
|
|
157
|
-
readonly name: "LLaMA 3.2 90B Vision (Preview)";
|
|
158
|
-
readonly description: "[BETA preview version] The Llama 3.2-Vision instruction-tuned models are optimized for visual recognition, image reasoning, captioning, and answering general questions about an image.";
|
|
159
|
-
readonly tags: readonly ["recommended", "vision", "general-purpose"];
|
|
160
|
-
readonly input: {
|
|
161
|
-
readonly costPer1MTokens: 0.9;
|
|
162
|
-
readonly maxTokens: 128000;
|
|
163
|
-
};
|
|
164
|
-
readonly output: {
|
|
165
|
-
readonly costPer1MTokens: 0.9;
|
|
166
|
-
readonly maxTokens: 8192;
|
|
167
|
-
};
|
|
168
|
-
readonly integration: "groq";
|
|
169
|
-
}, {
|
|
170
|
-
readonly id: "groq__mixtral-8x7b-32768";
|
|
171
|
-
readonly name: "Mixtral 8x7B";
|
|
172
|
-
readonly tags: readonly ["low-cost", "general-purpose", "deprecated"];
|
|
173
|
-
readonly description: "Mistral MoE 8x7B Instruct v0.1 model with Sparse Mixture of Experts. Fine tuned for instruction following";
|
|
174
|
-
readonly input: {
|
|
175
|
-
readonly costPer1MTokens: 0.24;
|
|
176
|
-
readonly maxTokens: 32768;
|
|
177
|
-
};
|
|
178
|
-
readonly output: {
|
|
179
|
-
readonly costPer1MTokens: 0.24;
|
|
180
|
-
readonly maxTokens: 32768;
|
|
181
|
-
};
|
|
182
|
-
readonly integration: "groq";
|
|
183
|
-
}, {
|
|
184
|
-
readonly id: "openai__gpt-3.5-turbo-0125";
|
|
185
|
-
readonly name: "GPT-3.5 Turbo";
|
|
186
|
-
readonly description: "GPT-3.5 Turbo can understand and generate natural language or code and has been optimized for chat but works well for non-chat tasks as well.";
|
|
187
|
-
readonly tags: readonly ["deprecated", "general-purpose", "low-cost"];
|
|
188
|
-
readonly input: {
|
|
189
|
-
readonly costPer1MTokens: 0.5;
|
|
190
|
-
readonly maxTokens: 128000;
|
|
191
|
-
};
|
|
192
|
-
readonly output: {
|
|
193
|
-
readonly costPer1MTokens: 1.5;
|
|
194
|
-
readonly maxTokens: 4096;
|
|
195
|
-
};
|
|
196
|
-
readonly integration: "openai";
|
|
197
|
-
}, {
|
|
198
|
-
readonly id: "openai__gpt-4-turbo-2024-04-09";
|
|
199
|
-
readonly name: "GPT-4 Turbo";
|
|
200
|
-
readonly description: "GPT-4 is a large multimodal model (accepting text or image inputs and outputting text) that can solve difficult problems with greater accuracy than any of our previous models, thanks to its broader general knowledge and advanced reasoning capabilities.";
|
|
201
|
-
readonly tags: readonly ["deprecated", "general-purpose", "coding", "agents", "function-calling"];
|
|
202
|
-
readonly input: {
|
|
203
|
-
readonly costPer1MTokens: 10;
|
|
204
|
-
readonly maxTokens: 128000;
|
|
205
|
-
};
|
|
206
|
-
readonly output: {
|
|
207
|
-
readonly costPer1MTokens: 30;
|
|
208
|
-
readonly maxTokens: 4096;
|
|
209
|
-
};
|
|
210
|
-
readonly integration: "openai";
|
|
211
|
-
}, {
|
|
212
|
-
readonly id: "openai__gpt-4o-2024-08-06";
|
|
213
|
-
readonly name: "GPT-4o (August 2024)";
|
|
214
|
-
readonly description: "GPT-4o (“o” for “omni”) is OpenAI's most advanced model. It is multimodal (accepting text or image inputs and outputting text), and it has the same high intelligence as GPT-4 Turbo but is cheaper and more efficient.";
|
|
215
|
-
readonly tags: readonly ["recommended", "vision", "general-purpose", "coding", "agents", "function-calling"];
|
|
216
|
-
readonly input: {
|
|
217
|
-
readonly costPer1MTokens: 2.5;
|
|
218
|
-
readonly maxTokens: 128000;
|
|
219
|
-
};
|
|
220
|
-
readonly output: {
|
|
221
|
-
readonly costPer1MTokens: 10;
|
|
222
|
-
readonly maxTokens: 16384;
|
|
223
|
-
};
|
|
224
|
-
readonly integration: "openai";
|
|
225
|
-
}, {
|
|
226
|
-
readonly id: "openai__gpt-4o-2024-05-13";
|
|
227
|
-
readonly name: "GPT-4o (May 2024)";
|
|
228
|
-
readonly description: "GPT-4o (“o” for “omni”) is OpenAI's most advanced model. It is multimodal (accepting text or image inputs and outputting text), and it has the same high intelligence as GPT-4 Turbo but is cheaper and more efficient.";
|
|
229
|
-
readonly tags: readonly ["vision", "general-purpose", "coding", "agents", "function-calling"];
|
|
230
|
-
readonly input: {
|
|
231
|
-
readonly costPer1MTokens: 5;
|
|
232
|
-
readonly maxTokens: 128000;
|
|
233
|
-
};
|
|
234
|
-
readonly output: {
|
|
235
|
-
readonly costPer1MTokens: 15;
|
|
236
|
-
readonly maxTokens: 4096;
|
|
237
|
-
};
|
|
238
|
-
readonly integration: "openai";
|
|
239
|
-
}, {
|
|
240
|
-
readonly id: "openai__gpt-4o-mini-2024-07-18";
|
|
241
|
-
readonly name: "GPT-4o Mini";
|
|
242
|
-
readonly description: "GPT-4o mini (“o” for “omni”) is OpenAI's most advanced model in the small models category, and their cheapest model yet. It is multimodal (accepting text or image inputs and outputting text), has higher intelligence than gpt-3.5-turbo but is just as fast. It is meant to be used for smaller tasks, including vision tasks. It's recommended to choose gpt-4o-mini where you would have previously used gpt-3.5-turbo as this model is more capable and cheaper.";
|
|
243
|
-
readonly tags: readonly ["recommended", "vision", "low-cost", "general-purpose", "function-calling"];
|
|
244
|
-
readonly input: {
|
|
245
|
-
readonly costPer1MTokens: 0.15;
|
|
246
|
-
readonly maxTokens: 128000;
|
|
247
|
-
};
|
|
248
|
-
readonly output: {
|
|
249
|
-
readonly costPer1MTokens: 0.6;
|
|
250
|
-
readonly maxTokens: 16384;
|
|
251
|
-
};
|
|
252
|
-
readonly integration: "openai";
|
|
253
|
-
}];
|
package/dist/src/models.js
DELETED
|
@@ -1,348 +0,0 @@
|
|
|
1
|
-
export const Models = [
|
|
2
|
-
{
|
|
3
|
-
"id": "anthropic__claude-3-haiku-20240307",
|
|
4
|
-
"name": "Claude 3 Haiku",
|
|
5
|
-
"description": "Claude 3 Haiku is Anthropic's fastest and most compact model for near-instant responsiveness. Quick and accurate targeted performance.",
|
|
6
|
-
"tags": [
|
|
7
|
-
"low-cost",
|
|
8
|
-
"general-purpose"
|
|
9
|
-
],
|
|
10
|
-
"input": {
|
|
11
|
-
"costPer1MTokens": 0.25,
|
|
12
|
-
"maxTokens": 200000
|
|
13
|
-
},
|
|
14
|
-
"output": {
|
|
15
|
-
"costPer1MTokens": 1.25,
|
|
16
|
-
"maxTokens": 4096
|
|
17
|
-
},
|
|
18
|
-
"integration": "anthropic"
|
|
19
|
-
},
|
|
20
|
-
{
|
|
21
|
-
"id": "anthropic__claude-3-5-sonnet-20240620",
|
|
22
|
-
"name": "Claude 3.5 Sonnet",
|
|
23
|
-
"description": "Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at coding, data science, visual processing, and agentic tasks.",
|
|
24
|
-
"tags": [
|
|
25
|
-
"recommended",
|
|
26
|
-
"vision",
|
|
27
|
-
"general-purpose",
|
|
28
|
-
"agents",
|
|
29
|
-
"coding",
|
|
30
|
-
"function-calling",
|
|
31
|
-
"storytelling"
|
|
32
|
-
],
|
|
33
|
-
"input": {
|
|
34
|
-
"costPer1MTokens": 3,
|
|
35
|
-
"maxTokens": 200000
|
|
36
|
-
},
|
|
37
|
-
"output": {
|
|
38
|
-
"costPer1MTokens": 15,
|
|
39
|
-
"maxTokens": 4096
|
|
40
|
-
},
|
|
41
|
-
"integration": "anthropic"
|
|
42
|
-
},
|
|
43
|
-
{
|
|
44
|
-
"id": "groq__gemma-7b-it",
|
|
45
|
-
"name": "Gemma 7B",
|
|
46
|
-
"description": "Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning.",
|
|
47
|
-
"tags": [
|
|
48
|
-
"deprecated",
|
|
49
|
-
"low-cost"
|
|
50
|
-
],
|
|
51
|
-
"input": {
|
|
52
|
-
"costPer1MTokens": 0.07,
|
|
53
|
-
"maxTokens": 8192
|
|
54
|
-
},
|
|
55
|
-
"output": {
|
|
56
|
-
"costPer1MTokens": 0.07,
|
|
57
|
-
"maxTokens": 8192
|
|
58
|
-
},
|
|
59
|
-
"integration": "groq"
|
|
60
|
-
},
|
|
61
|
-
{
|
|
62
|
-
"id": "groq__gemma2-9b-it",
|
|
63
|
-
"name": "Gemma2 9B",
|
|
64
|
-
"description": "Redesigned for outsized performance and unmatched efficiency, Gemma 2 optimizes for blazing-fast inference on diverse hardware. Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. They are text-to-text, decoder-only large language models, available in English, with open weights, pre-trained variants, and instruction-tuned variants. Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning.",
|
|
65
|
-
"tags": [
|
|
66
|
-
"low-cost",
|
|
67
|
-
"general-purpose"
|
|
68
|
-
],
|
|
69
|
-
"input": {
|
|
70
|
-
"costPer1MTokens": 0.2,
|
|
71
|
-
"maxTokens": 8192
|
|
72
|
-
},
|
|
73
|
-
"output": {
|
|
74
|
-
"costPer1MTokens": 0.2,
|
|
75
|
-
"maxTokens": 8192
|
|
76
|
-
},
|
|
77
|
-
"integration": "groq"
|
|
78
|
-
},
|
|
79
|
-
{
|
|
80
|
-
"id": "groq__llama3-70b-8192",
|
|
81
|
-
"name": "LLaMA 3 70B",
|
|
82
|
-
"tags": [
|
|
83
|
-
"general-purpose"
|
|
84
|
-
],
|
|
85
|
-
"description": "Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. The Llama 3 instruction tuned models are optimized for dialogue use cases and outperform many of the available open source chat models on common industry benchmarks.",
|
|
86
|
-
"input": {
|
|
87
|
-
"costPer1MTokens": 0.59,
|
|
88
|
-
"maxTokens": 8192
|
|
89
|
-
},
|
|
90
|
-
"output": {
|
|
91
|
-
"costPer1MTokens": 0.79,
|
|
92
|
-
"maxTokens": 8192
|
|
93
|
-
},
|
|
94
|
-
"integration": "groq"
|
|
95
|
-
},
|
|
96
|
-
{
|
|
97
|
-
"id": "groq__llama3-8b-8192",
|
|
98
|
-
"name": "LLaMA 3 8B",
|
|
99
|
-
"description": "Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. The Llama 3 instruction tuned models are optimized for dialogue use cases and outperform many of the available open source chat models on common industry benchmarks.",
|
|
100
|
-
"tags": [
|
|
101
|
-
"low-cost",
|
|
102
|
-
"general-purpose",
|
|
103
|
-
"deprecated"
|
|
104
|
-
],
|
|
105
|
-
"input": {
|
|
106
|
-
"costPer1MTokens": 0.05,
|
|
107
|
-
"maxTokens": 8192
|
|
108
|
-
},
|
|
109
|
-
"output": {
|
|
110
|
-
"costPer1MTokens": 0.08,
|
|
111
|
-
"maxTokens": 8192
|
|
112
|
-
},
|
|
113
|
-
"integration": "groq"
|
|
114
|
-
},
|
|
115
|
-
{
|
|
116
|
-
"id": "groq__llama-3.1-70b-versatile",
|
|
117
|
-
"name": "LLaMA 3.1 70B",
|
|
118
|
-
"description": "The Llama 3.1 instruction-tuned, text-only models are optimized for multilingual dialogue use cases.",
|
|
119
|
-
"tags": [
|
|
120
|
-
"recommended",
|
|
121
|
-
"general-purpose"
|
|
122
|
-
],
|
|
123
|
-
"input": {
|
|
124
|
-
"costPer1MTokens": 0.59,
|
|
125
|
-
"maxTokens": 128000
|
|
126
|
-
},
|
|
127
|
-
"output": {
|
|
128
|
-
"costPer1MTokens": 0.79,
|
|
129
|
-
"maxTokens": 8192
|
|
130
|
-
},
|
|
131
|
-
"integration": "groq"
|
|
132
|
-
},
|
|
133
|
-
{
|
|
134
|
-
"id": "groq__llama-3.1-8b-instant",
|
|
135
|
-
"name": "LLaMA 3.1 8B",
|
|
136
|
-
"description": "The Llama 3.1 instruction-tuned, text-only models are optimized for multilingual dialogue use cases.",
|
|
137
|
-
"tags": [
|
|
138
|
-
"low-cost",
|
|
139
|
-
"general-purpose"
|
|
140
|
-
],
|
|
141
|
-
"input": {
|
|
142
|
-
"costPer1MTokens": 0.05,
|
|
143
|
-
"maxTokens": 128000
|
|
144
|
-
},
|
|
145
|
-
"output": {
|
|
146
|
-
"costPer1MTokens": 0.08,
|
|
147
|
-
"maxTokens": 8192
|
|
148
|
-
},
|
|
149
|
-
"integration": "groq"
|
|
150
|
-
},
|
|
151
|
-
{
|
|
152
|
-
"id": "groq__llama-3.2-11b-vision-preview",
|
|
153
|
-
"name": "LLaMA 3.2 11B Vision (Preview)",
|
|
154
|
-
"description": "[BETA preview version] The Llama 3.2-Vision instruction-tuned models are optimized for visual recognition, image reasoning, captioning, and answering general questions about an image.",
|
|
155
|
-
"tags": [
|
|
156
|
-
"low-cost",
|
|
157
|
-
"vision",
|
|
158
|
-
"general-purpose"
|
|
159
|
-
],
|
|
160
|
-
"input": {
|
|
161
|
-
"costPer1MTokens": 0.18,
|
|
162
|
-
"maxTokens": 128000
|
|
163
|
-
},
|
|
164
|
-
"output": {
|
|
165
|
-
"costPer1MTokens": 0.18,
|
|
166
|
-
"maxTokens": 8192
|
|
167
|
-
},
|
|
168
|
-
"integration": "groq"
|
|
169
|
-
},
|
|
170
|
-
{
|
|
171
|
-
"id": "groq__llama-3.2-1b-preview",
|
|
172
|
-
"name": "LLaMA 3.2 1B (Preview)",
|
|
173
|
-
"description": "[BETA preview version] The Llama 3.2 instruction-tuned, text-only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.",
|
|
174
|
-
"tags": [
|
|
175
|
-
"low-cost"
|
|
176
|
-
],
|
|
177
|
-
"input": {
|
|
178
|
-
"costPer1MTokens": 0.04,
|
|
179
|
-
"maxTokens": 128000
|
|
180
|
-
},
|
|
181
|
-
"output": {
|
|
182
|
-
"costPer1MTokens": 0.04,
|
|
183
|
-
"maxTokens": 8192
|
|
184
|
-
},
|
|
185
|
-
"integration": "groq"
|
|
186
|
-
},
|
|
187
|
-
{
|
|
188
|
-
"id": "groq__llama-3.2-3b-preview",
|
|
189
|
-
"name": "LLaMA 3.2 3B (Preview)",
|
|
190
|
-
"description": "[BETA preview version] The Llama 3.2 instruction-tuned, text-only models are optimized for multilingual dialogue use cases, including agentic retrieval and summarization tasks.",
|
|
191
|
-
"tags": [
|
|
192
|
-
"low-cost",
|
|
193
|
-
"general-purpose"
|
|
194
|
-
],
|
|
195
|
-
"input": {
|
|
196
|
-
"costPer1MTokens": 0.06,
|
|
197
|
-
"maxTokens": 128000
|
|
198
|
-
},
|
|
199
|
-
"output": {
|
|
200
|
-
"costPer1MTokens": 0.06,
|
|
201
|
-
"maxTokens": 8192
|
|
202
|
-
},
|
|
203
|
-
"integration": "groq"
|
|
204
|
-
},
|
|
205
|
-
{
|
|
206
|
-
"id": "groq__llama-3.2-90b-vision-preview",
|
|
207
|
-
"name": "LLaMA 3.2 90B Vision (Preview)",
|
|
208
|
-
"description": "[BETA preview version] The Llama 3.2-Vision instruction-tuned models are optimized for visual recognition, image reasoning, captioning, and answering general questions about an image.",
|
|
209
|
-
"tags": [
|
|
210
|
-
"recommended",
|
|
211
|
-
"vision",
|
|
212
|
-
"general-purpose"
|
|
213
|
-
],
|
|
214
|
-
"input": {
|
|
215
|
-
"costPer1MTokens": 0.9,
|
|
216
|
-
"maxTokens": 128000
|
|
217
|
-
},
|
|
218
|
-
"output": {
|
|
219
|
-
"costPer1MTokens": 0.9,
|
|
220
|
-
"maxTokens": 8192
|
|
221
|
-
},
|
|
222
|
-
"integration": "groq"
|
|
223
|
-
},
|
|
224
|
-
{
|
|
225
|
-
"id": "groq__mixtral-8x7b-32768",
|
|
226
|
-
"name": "Mixtral 8x7B",
|
|
227
|
-
"tags": [
|
|
228
|
-
"low-cost",
|
|
229
|
-
"general-purpose",
|
|
230
|
-
"deprecated"
|
|
231
|
-
],
|
|
232
|
-
"description": "Mistral MoE 8x7B Instruct v0.1 model with Sparse Mixture of Experts. Fine tuned for instruction following",
|
|
233
|
-
"input": {
|
|
234
|
-
"costPer1MTokens": 0.24,
|
|
235
|
-
"maxTokens": 32768
|
|
236
|
-
},
|
|
237
|
-
"output": {
|
|
238
|
-
"costPer1MTokens": 0.24,
|
|
239
|
-
"maxTokens": 32768
|
|
240
|
-
},
|
|
241
|
-
"integration": "groq"
|
|
242
|
-
},
|
|
243
|
-
{
|
|
244
|
-
"id": "openai__gpt-3.5-turbo-0125",
|
|
245
|
-
"name": "GPT-3.5 Turbo",
|
|
246
|
-
"description": "GPT-3.5 Turbo can understand and generate natural language or code and has been optimized for chat but works well for non-chat tasks as well.",
|
|
247
|
-
"tags": [
|
|
248
|
-
"deprecated",
|
|
249
|
-
"general-purpose",
|
|
250
|
-
"low-cost"
|
|
251
|
-
],
|
|
252
|
-
"input": {
|
|
253
|
-
"costPer1MTokens": 0.5,
|
|
254
|
-
"maxTokens": 128000
|
|
255
|
-
},
|
|
256
|
-
"output": {
|
|
257
|
-
"costPer1MTokens": 1.5,
|
|
258
|
-
"maxTokens": 4096
|
|
259
|
-
},
|
|
260
|
-
"integration": "openai"
|
|
261
|
-
},
|
|
262
|
-
{
|
|
263
|
-
"id": "openai__gpt-4-turbo-2024-04-09",
|
|
264
|
-
"name": "GPT-4 Turbo",
|
|
265
|
-
"description": "GPT-4 is a large multimodal model (accepting text or image inputs and outputting text) that can solve difficult problems with greater accuracy than any of our previous models, thanks to its broader general knowledge and advanced reasoning capabilities.",
|
|
266
|
-
"tags": [
|
|
267
|
-
"deprecated",
|
|
268
|
-
"general-purpose",
|
|
269
|
-
"coding",
|
|
270
|
-
"agents",
|
|
271
|
-
"function-calling"
|
|
272
|
-
],
|
|
273
|
-
"input": {
|
|
274
|
-
"costPer1MTokens": 10,
|
|
275
|
-
"maxTokens": 128000
|
|
276
|
-
},
|
|
277
|
-
"output": {
|
|
278
|
-
"costPer1MTokens": 30,
|
|
279
|
-
"maxTokens": 4096
|
|
280
|
-
},
|
|
281
|
-
"integration": "openai"
|
|
282
|
-
},
|
|
283
|
-
{
|
|
284
|
-
"id": "openai__gpt-4o-2024-08-06",
|
|
285
|
-
"name": "GPT-4o (August 2024)",
|
|
286
|
-
"description": "GPT-4o (“o” for “omni”) is OpenAI's most advanced model. It is multimodal (accepting text or image inputs and outputting text), and it has the same high intelligence as GPT-4 Turbo but is cheaper and more efficient.",
|
|
287
|
-
"tags": [
|
|
288
|
-
"recommended",
|
|
289
|
-
"vision",
|
|
290
|
-
"general-purpose",
|
|
291
|
-
"coding",
|
|
292
|
-
"agents",
|
|
293
|
-
"function-calling"
|
|
294
|
-
],
|
|
295
|
-
"input": {
|
|
296
|
-
"costPer1MTokens": 2.5,
|
|
297
|
-
"maxTokens": 128000
|
|
298
|
-
},
|
|
299
|
-
"output": {
|
|
300
|
-
"costPer1MTokens": 10,
|
|
301
|
-
"maxTokens": 16384
|
|
302
|
-
},
|
|
303
|
-
"integration": "openai"
|
|
304
|
-
},
|
|
305
|
-
{
|
|
306
|
-
"id": "openai__gpt-4o-2024-05-13",
|
|
307
|
-
"name": "GPT-4o (May 2024)",
|
|
308
|
-
"description": "GPT-4o (“o” for “omni”) is OpenAI's most advanced model. It is multimodal (accepting text or image inputs and outputting text), and it has the same high intelligence as GPT-4 Turbo but is cheaper and more efficient.",
|
|
309
|
-
"tags": [
|
|
310
|
-
"vision",
|
|
311
|
-
"general-purpose",
|
|
312
|
-
"coding",
|
|
313
|
-
"agents",
|
|
314
|
-
"function-calling"
|
|
315
|
-
],
|
|
316
|
-
"input": {
|
|
317
|
-
"costPer1MTokens": 5,
|
|
318
|
-
"maxTokens": 128000
|
|
319
|
-
},
|
|
320
|
-
"output": {
|
|
321
|
-
"costPer1MTokens": 15,
|
|
322
|
-
"maxTokens": 4096
|
|
323
|
-
},
|
|
324
|
-
"integration": "openai"
|
|
325
|
-
},
|
|
326
|
-
{
|
|
327
|
-
"id": "openai__gpt-4o-mini-2024-07-18",
|
|
328
|
-
"name": "GPT-4o Mini",
|
|
329
|
-
"description": "GPT-4o mini (“o” for “omni”) is OpenAI's most advanced model in the small models category, and their cheapest model yet. It is multimodal (accepting text or image inputs and outputting text), has higher intelligence than gpt-3.5-turbo but is just as fast. It is meant to be used for smaller tasks, including vision tasks. It's recommended to choose gpt-4o-mini where you would have previously used gpt-3.5-turbo as this model is more capable and cheaper.",
|
|
330
|
-
"tags": [
|
|
331
|
-
"recommended",
|
|
332
|
-
"vision",
|
|
333
|
-
"low-cost",
|
|
334
|
-
"general-purpose",
|
|
335
|
-
"function-calling"
|
|
336
|
-
],
|
|
337
|
-
"input": {
|
|
338
|
-
"costPer1MTokens": 0.15,
|
|
339
|
-
"maxTokens": 128000
|
|
340
|
-
},
|
|
341
|
-
"output": {
|
|
342
|
-
"costPer1MTokens": 0.6,
|
|
343
|
-
"maxTokens": 16384
|
|
344
|
-
},
|
|
345
|
-
"integration": "openai"
|
|
346
|
-
}
|
|
347
|
-
];
|
|
348
|
-
//# sourceMappingURL=models.js.map
|
package/dist/src/models.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"models.js","sourceRoot":"","sources":["../../src/models.ts"],"names":[],"mappings":"AAAA,MAAM,CAAC,MAAM,MAAM,GAAG;IACpB;QACE,IAAI,EAAE,oCAAoC;QAC1C,MAAM,EAAE,gBAAgB;QACxB,aAAa,EAAE,wIAAwI;QACvJ,MAAM,EAAE;YACN,UAAU;YACV,iBAAiB;SAClB;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,WAAW;KAC3B;IACD;QACE,IAAI,EAAE,uCAAuC;QAC7C,MAAM,EAAE,mBAAmB;QAC3B,aAAa,EAAE,4MAA4M;QAC3N,MAAM,EAAE;YACN,aAAa;YACb,QAAQ;YACR,iBAAiB;YACjB,QAAQ;YACR,QAAQ;YACR,kBAAkB;YAClB,cAAc;SACf;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,CAAC;YACpB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,EAAE;YACrB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,WAAW;KAC3B;IACD;QACE,IAAI,EAAE,mBAAmB;QACzB,MAAM,EAAE,UAAU;QAClB,aAAa,EAAE,sbAAsb;QACrc,MAAM,EAAE;YACN,YAAY;YACZ,UAAU;SACX;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,IAAI;SAClB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,MAAM;KACtB;IACD;QACE,IAAI,EAAE,oBAAoB;QAC1B,MAAM,EAAE,WAAW;QACnB,aAAa,EAAE,sjBAAsjB;QACrkB,MAAM,EAAE;YACN,UAAU;YACV,iBAAiB;SAClB;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,GAAG;YACtB,WAAW,EAAE,IAAI;SAClB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,GAAG;YACtB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,MAAM;KACtB;IACD;QACE,IAAI,EAAE,uBAAuB;QAC7B,MAAM,EAAE,aAAa;QACrB,MAAM,EAAE;YACN,iBAAiB;SAClB;QACD,aAAa,EAAE,wVAAwV;QACvW,OAAO,EAAE;YACP,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,IAAI;SAClB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,MAAM;KACtB;IACD;QACE,IAAI,EAAE,sBAAsB;QAC5B,MAAM,EAAE,YAAY;QACpB,aAAa,EAAE,wVAAwV;QACvW,MAAM,EAAE;YACN,UAAU;YACV,iBAAiB;YACjB,YAAY;SACb;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,IAAI;SAClB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,MAAM;KACtB;IACD;QACE,IAAI,EAAE,+BAA+B;QACrC,MAAM,EAAE,eAAe;QACvB,aAAa,EAAE,sGAAsG;QACrH,MAAM,EAAE;YACN,aAAa;YACb,iBAAiB;SAClB;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,MAAM;KACtB;IACD;QACE,IAAI,EAAE,4BAA4B;QAClC,MAAM,EAAE,cAAc;QACtB,aAAa,EAAE,sGAAsG;QACrH,MAAM,EAAE;YACN,UAAU;YACV,iBAAiB;SAClB;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,MAAM;KACtB;IACD;QACE,IAAI,EAAE,oCAAoC;QAC1C,MAAM,EAAE,gCAAgC;QACxC,aAAa,EAAE,yLAAyL;QACxM,MAAM,EAAE;YACN,UAAU;YACV,QAAQ;YACR,iBAAiB;SAClB;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,MAAM;KACtB;IACD;QACE,IAAI,EAAE,4BAA4B;QAClC,MAAM,EAAE,wBAAwB;QAChC,aAAa,EAAE,kLAAkL;QACjM,MAAM,EAAE;YACN,UAAU;SACX;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,MAAM;KACtB;IACD;QACE,IAAI,EAAE,4BAA4B;QAClC,MAAM,EAAE,wBAAwB;QAChC,aAAa,EAAE,kLAAkL;QACjM,MAAM,EAAE;YACN,UAAU;YACV,iBAAiB;SAClB;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,MAAM;KACtB;IACD;QACE,IAAI,EAAE,oCAAoC;QAC1C,MAAM,EAAE,gCAAgC;QACxC,aAAa,EAAE,yLAAyL;QACxM,MAAM,EAAE;YACN,aAAa;YACb,QAAQ;YACR,iBAAiB;SAClB;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,GAAG;YACtB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,GAAG;YACtB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,MAAM;KACtB;IACD;QACE,IAAI,EAAE,0BAA0B;QAChC,MAAM,EAAE,cAAc;QACtB,MAAM,EAAE;YACN,UAAU;YACV,iBAAiB;YACjB,YAAY;SACb;QACD,aAAa,EAAE,2GAA2G;QAC1H,OAAO,EAAE;YACP,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,KAAK;SACnB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,KAAK;SACnB;QACD,aAAa,EAAE,MAAM;KACtB;IACD;QACE,IAAI,EAAE,4BAA4B;QAClC,MAAM,EAAE,eAAe;QACvB,aAAa,EAAE,+IAA+I;QAC9J,MAAM,EAAE;YACN,YAAY;YACZ,iBAAiB;YACjB,UAAU;SACX;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,GAAG;YACtB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,GAAG;YACtB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,QAAQ;KACxB;IACD;QACE,IAAI,EAAE,gCAAgC;QACtC,MAAM,EAAE,aAAa;QACrB,aAAa,EAAE,8PAA8P;QAC7Q,MAAM,EAAE;YACN,YAAY;YACZ,iBAAiB;YACjB,QAAQ;YACR,QAAQ;YACR,kBAAkB;SACnB;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,EAAE;YACrB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,EAAE;YACrB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,QAAQ;KACxB;IACD;QACE,IAAI,EAAE,2BAA2B;QACjC,MAAM,EAAE,sBAAsB;QAC9B,aAAa,EAAE,yNAAyN;QACxO,MAAM,EAAE;YACN,aAAa;YACb,QAAQ;YACR,iBAAiB;YACjB,QAAQ;YACR,QAAQ;YACR,kBAAkB;SACnB;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,GAAG;YACtB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,EAAE;YACrB,WAAW,EAAE,KAAK;SACnB;QACD,aAAa,EAAE,QAAQ;KACxB;IACD;QACE,IAAI,EAAE,2BAA2B;QACjC,MAAM,EAAE,mBAAmB;QAC3B,aAAa,EAAE,yNAAyN;QACxO,MAAM,EAAE;YACN,QAAQ;YACR,iBAAiB;YACjB,QAAQ;YACR,QAAQ;YACR,kBAAkB;SACnB;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,CAAC;YACpB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,EAAE;YACrB,WAAW,EAAE,IAAI;SAClB;QACD,aAAa,EAAE,QAAQ;KACxB;IACD;QACE,IAAI,EAAE,gCAAgC;QACtC,MAAM,EAAE,aAAa;QACrB,aAAa,EAAE,wcAAwc;QACvd,MAAM,EAAE;YACN,aAAa;YACb,QAAQ;YACR,UAAU;YACV,iBAAiB;YACjB,kBAAkB;SACnB;QACD,OAAO,EAAE;YACP,iBAAiB,EAAE,IAAI;YACvB,WAAW,EAAE,MAAM;SACpB;QACD,QAAQ,EAAE;YACR,iBAAiB,EAAE,GAAG;YACtB,WAAW,EAAE,KAAK;SACnB;QACD,aAAa,EAAE,QAAQ;KACxB;CACO,CAAA"}
|
|
@@ -1,19 +0,0 @@
|
|
|
1
|
-
import { Client } from '@botpress/client';
|
|
2
|
-
import { type TextTokenizer } from '@botpress/wasm';
|
|
3
|
-
import { Zai } from '../..';
|
|
4
|
-
export declare const getClient: () => Client;
|
|
5
|
-
export declare const getZai: () => Zai;
|
|
6
|
-
export declare let tokenizer: TextTokenizer;
|
|
7
|
-
export declare const BotpressDocumentation: string;
|
|
8
|
-
export declare const metadata: {
|
|
9
|
-
cost: {
|
|
10
|
-
input: number;
|
|
11
|
-
output: number;
|
|
12
|
-
};
|
|
13
|
-
latency: number;
|
|
14
|
-
model: string;
|
|
15
|
-
tokens: {
|
|
16
|
-
input: number;
|
|
17
|
-
output: number;
|
|
18
|
-
};
|
|
19
|
-
};
|
|
@@ -1,24 +0,0 @@
|
|
|
1
|
-
import { Client } from '@botpress/client';
|
|
2
|
-
import { getWasmTokenizer } from '@botpress/wasm';
|
|
3
|
-
import fs from 'node:fs';
|
|
4
|
-
import path from 'node:path';
|
|
5
|
-
import { beforeAll } from 'vitest';
|
|
6
|
-
import { Zai } from '../..';
|
|
7
|
-
export const getClient = () => {
|
|
8
|
-
return new Client({
|
|
9
|
-
apiUrl: process.env.VITE_CLOUD_API_ENDPOINT,
|
|
10
|
-
botId: process.env.VITEST_BOT_ID ?? process.env.VITE_BOT_ID,
|
|
11
|
-
token: process.env.VITEST_CLOUD_PAT ?? process.env.VITE_CLOUD_PAT
|
|
12
|
-
});
|
|
13
|
-
};
|
|
14
|
-
export const getZai = () => {
|
|
15
|
-
const client = getClient();
|
|
16
|
-
return new Zai({ client, retry: { maxRetries: 0 } });
|
|
17
|
-
};
|
|
18
|
-
export let tokenizer = null;
|
|
19
|
-
beforeAll(async () => {
|
|
20
|
-
tokenizer = await getWasmTokenizer();
|
|
21
|
-
});
|
|
22
|
-
export const BotpressDocumentation = fs.readFileSync(path.join(__dirname, './botpress_docs.txt'), 'utf-8').trim();
|
|
23
|
-
export const metadata = { cost: { input: 1, output: 1 }, latency: 0, model: '', tokens: { input: 1, output: 1 } };
|
|
24
|
-
//# sourceMappingURL=index.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"index.js","sourceRoot":"","sources":["../../../../src/operations/__tests/index.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,MAAM,EAAE,MAAM,kBAAkB,CAAA;AACzC,OAAO,EAAsB,gBAAgB,EAAE,MAAM,gBAAgB,CAAA;AACrE,OAAO,EAAE,MAAM,SAAS,CAAA;AACxB,OAAO,IAAI,MAAM,WAAW,CAAA;AAC5B,OAAO,EAAE,SAAS,EAAE,MAAM,QAAQ,CAAA;AAElC,OAAO,EAAE,GAAG,EAAE,MAAM,OAAO,CAAA;AAE3B,MAAM,CAAC,MAAM,SAAS,GAAG,GAAG,EAAE;IAC5B,OAAO,IAAI,MAAM,CAAC;QAChB,MAAM,EAAE,OAAO,CAAC,GAAG,CAAC,uBAAuB;QAC3C,KAAK,EAAE,OAAO,CAAC,GAAG,CAAC,aAAa,IAAI,OAAO,CAAC,GAAG,CAAC,WAAW;QAC3D,KAAK,EAAE,OAAO,CAAC,GAAG,CAAC,gBAAgB,IAAI,OAAO,CAAC,GAAG,CAAC,cAAc;KAClE,CAAC,CAAA;AACJ,CAAC,CAAA;AAED,MAAM,CAAC,MAAM,MAAM,GAAG,GAAG,EAAE;IACzB,MAAM,MAAM,GAAG,SAAS,EAAE,CAAA;IAC1B,OAAO,IAAI,GAAG,CAAC,EAAE,MAAM,EAAE,KAAK,EAAE,EAAE,UAAU,EAAE,CAAC,EAAE,EAAE,CAAC,CAAA;AACtD,CAAC,CAAA;AAED,MAAM,CAAC,IAAI,SAAS,GAAkB,IAAK,CAAA;AAE3C,SAAS,CAAC,KAAK,IAAI,EAAE;IACnB,SAAS,GAAG,MAAM,gBAAgB,EAAE,CAAA;AACtC,CAAC,CAAC,CAAA;AAEF,MAAM,CAAC,MAAM,qBAAqB,GAAG,EAAE,CAAC,YAAY,CAAC,IAAI,CAAC,IAAI,CAAC,SAAS,EAAE,qBAAqB,CAAC,EAAE,OAAO,CAAC,CAAC,IAAI,EAAE,CAAA;AAEjH,MAAM,CAAC,MAAM,QAAQ,GAAG,EAAE,IAAI,EAAE,EAAE,KAAK,EAAE,CAAC,EAAE,MAAM,EAAE,CAAC,EAAE,EAAE,OAAO,EAAE,CAAC,EAAE,KAAK,EAAE,EAAE,EAAE,MAAM,EAAE,EAAE,KAAK,EAAE,CAAC,EAAE,MAAM,EAAE,CAAC,EAAE,EAAE,CAAA"}
|