modprompt 0.9.2 → 0.9.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/db.d.ts CHANGED
@@ -1,3 +1,3 @@
1
- import { LmTemplate } from "./interfaces";
1
+ import { LmTemplate } from "./interfaces.js";
2
2
  declare const templates: Record<string, LmTemplate>;
3
3
  export { templates };
package/dist/main.js CHANGED
@@ -1,357 +1,411 @@
1
1
  const templates = {
2
- "none": {
3
- "id": "none",
4
- "name": "No template",
5
- "user": "{prompt}",
6
- "assistant": "",
7
- },
8
2
  "alpaca": {
3
+ "assistant": "### Response:",
9
4
  "id": "alpaca",
5
+ "linebreaks": {
6
+ "system": 2,
7
+ "user": 2
8
+ },
10
9
  "name": "Alpaca",
11
10
  "system": {
12
- "schema": "{system}",
13
11
  "message": "Below is an instruction that describes a task. Write a response that appropriately completes the request.",
12
+ "schema": "{system}"
14
13
  },
15
- "user": "### Instruction:\n{prompt}",
16
- "assistant": "### Response:",
17
- "linebreaks": {
18
- "system": 2,
19
- "user": 2
20
- }
14
+ "user": "### Instruction:\n{prompt}"
21
15
  },
22
- "llama": {
23
- "id": "llama",
24
- "name": "Llama",
16
+ "cerebrum": {
17
+ "assistant": "Ai:",
18
+ "id": "cerebrum",
19
+ "linebreaks": {
20
+ "user": 1
21
+ },
22
+ "name": "Cerebrum",
23
+ "prefix": "<s>",
24
+ "stop": [
25
+ "</s>"
26
+ ],
25
27
  "system": {
26
- "schema": "[INST] <<SYS>>\n{system}\n<</SYS>>",
27
- "message": "You are a helpful, respectful and honest assistant. Always answer as helpfully as possible\n\nIf a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information."
28
+ "message": "A chat between a user and a thinking artificial intelligence assistant. The assistant describes its thought process and gives helpful and detailed answers to the user's questions.",
29
+ "schema": "{system}"
28
30
  },
29
- "user": "{prompt}",
30
- "assistant": " [/INST] ",
31
+ "user": "User: {prompt}"
32
+ },
33
+ "chatml": {
34
+ "afterShot": " <|im_end|>\n",
35
+ "assistant": "<|im_start|>assistant",
36
+ "id": "chatml",
31
37
  "linebreaks": {
32
- "system": 2,
33
- "user": 0
38
+ "assistant": 1,
39
+ "system": 1,
40
+ "user": 1
34
41
  },
35
- "prefix": "<s>",
36
- "stop": ["</s>"],
42
+ "name": "ChatMl",
43
+ "stop": [
44
+ "<|im_end|>"
45
+ ],
46
+ "system": {
47
+ "schema": "<|im_start|>system\n{system}<|im_end|>"
48
+ },
49
+ "user": "<|im_start|>user\n{prompt}<|im_end|>"
37
50
  },
38
- "mistral": {
39
- "id": "mistral",
40
- "name": "Mistral",
41
- "user": "[INST] {prompt}",
51
+ "codestral": {
52
+ "afterShot": "\n",
42
53
  "assistant": " [/INST]",
43
- "stop": ["</s>"],
44
- "afterShot": "\n"
45
- },
46
- "orca": {
47
- "id": "orca",
48
- "name": "Orca",
54
+ "id": "codestral",
55
+ "linebreaks": {
56
+ "system": 2
57
+ },
58
+ "name": "Codestral",
59
+ "stop": [
60
+ "</s>"
61
+ ],
49
62
  "system": {
50
- "schema": "### System:\n{system}",
51
- "message": "You are an AI assistant that follows instruction extremely well. Help as much as you can.",
63
+ "schema": "<<SYS>>\n{system}\n<</SYS>>"
52
64
  },
53
- "user": "### User:\n{prompt}",
54
- "assistant": "### Response:",
65
+ "user": "[INST] {prompt}"
66
+ },
67
+ "command-r": {
68
+ "assistant": "<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>",
69
+ "id": "command-r",
55
70
  "linebreaks": {
56
- "system": 2,
57
- "user": 2
71
+ "user": 1
58
72
  },
73
+ "name": "Command-R",
74
+ "prefix": "<BOS_TOKEN>",
75
+ "stop": [
76
+ "<|END_OF_TURN_TOKEN|>"
77
+ ],
78
+ "system": {
79
+ "schema": "<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>{system}<|END_OF_TURN_TOKEN|>"
80
+ },
81
+ "user": "<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{prompt}<|END_OF_TURN_TOKEN|>"
59
82
  },
60
- "vicuna": {
61
- "id": "vicuna",
62
- "name": "Vicuna",
63
- "user": "USER: {prompt}",
64
- "assistant": "### ASSISTANT:",
83
+ "deepseek": {
84
+ "afterShot": "\n",
85
+ "assistant": "### Response:",
86
+ "id": "deepseek",
65
87
  "linebreaks": {
66
- "user": 2
88
+ "system": 1,
89
+ "user": 1
67
90
  },
68
- },
69
- "vicuna_system": {
70
- "id": "vicuna_system",
71
- "name": "Vicuna system",
91
+ "name": "Deepseek",
92
+ "stop": [
93
+ "<|EOT|>",
94
+ "### Instruction:"
95
+ ],
72
96
  "system": {
73
- "schema": "SYSTEM: {system}",
97
+ "message": "You are an AI programming assistant, utilizing the DeepSeek Coder model, developed by DeepSeek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer.",
98
+ "schema": "{system}"
74
99
  },
75
- "user": "USER: {prompt}",
76
- "assistant": "### ASSISTANT:",
100
+ "user": "### Instruction:\n{prompt}"
101
+ },
102
+ "deepseek2": {
103
+ "assistant": "Assistant:",
104
+ "id": "deepseek2",
77
105
  "linebreaks": {
78
106
  "system": 2,
79
107
  "user": 2
80
108
  },
81
- },
82
- "wizard_vicuna": {
83
- "id": "wizard_vicuna",
84
- "name": "Wizard Vicuna",
85
- "user": "### Human:\n{prompt}",
86
- "assistant": "### ASSISTANT:",
87
- "linebreaks": {
88
- "user": 2
109
+ "name": "Deepseek v2",
110
+ "stop": [
111
+ "<|end▁of▁sentence|>",
112
+ "<|tool▁calls▁end|>"
113
+ ],
114
+ "system": {
115
+ "schema": "<|begin▁of▁sentence|>{system}"
89
116
  },
90
- "stop": ["<|endoftext|>"]
117
+ "user": "User: {prompt}"
118
+ },
119
+ "gemma": {
120
+ "afterShot": "\n",
121
+ "assistant": "<end_of_turn>\n<start_of_turn>model",
122
+ "id": "gemma",
123
+ "name": "Gemma",
124
+ "stop": [
125
+ "<end_of_turn>"
126
+ ],
127
+ "user": "<start_of_turn>user\n{prompt}"
91
128
  },
92
129
  "guanaco": {
93
- "id": "guanaco",
94
- "name": "Guanaco",
95
- "user": "### Human: {prompt}",
96
130
  "assistant": "### Assistant:",
131
+ "id": "guanaco",
97
132
  "linebreaks": {
98
133
  "user": 1
99
134
  },
135
+ "name": "Guanaco",
136
+ "user": "### Human: {prompt}"
100
137
  },
101
- "chatml": {
102
- "id": "chatml",
103
- "name": "ChatMl",
104
- "system": {
105
- "schema": "<|im_start|>system\n{system}\n<|im_end|>",
106
- },
107
- "user": "<|im_start|>user\n{prompt}<|im_end|>",
108
- "assistant": "<|im_start|>assistant",
138
+ "human_response": {
139
+ "assistant": "### RESPONSE:",
140
+ "id": "human_response",
109
141
  "linebreaks": {
110
- "system": 1,
111
- "user": 1,
112
142
  "assistant": 1,
143
+ "user": 2
113
144
  },
114
- "stop": ["<|im_end|>"],
115
- "afterShot": " <|im_end|>",
145
+ "name": "Human response",
146
+ "user": "### HUMAN:\n{prompt}"
116
147
  },
117
- "zephyr": {
118
- "id": "zephyr",
119
- "name": "Zephyr",
120
- "system": {
121
- "schema": "<|system|>\n{system}<|endoftext|>",
122
- },
123
- "user": "<|user|>\n{prompt}<|endoftext|>",
124
- "assistant": "<|assistant|>",
148
+ "llama": {
149
+ "assistant": " [/INST] ",
150
+ "id": "llama",
125
151
  "linebreaks": {
126
- "system": 1,
127
- "user": 1,
128
- "assistant": 1,
152
+ "system": 2,
153
+ "user": 0
129
154
  },
130
- "afterShot": "\n",
131
- "stop": ["<|endoftext|>"]
155
+ "name": "Llama",
156
+ "prefix": "<s>",
157
+ "stop": [
158
+ "</s>"
159
+ ],
160
+ "system": {
161
+ "message": "You are a helpful, respectful and honest assistant. Always answer as helpfully as possible\n\nIf a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.",
162
+ "schema": "[INST] <<SYS>>\n{system}\n<</SYS>>"
163
+ },
164
+ "user": "{prompt}"
132
165
  },
133
- "octopus": {
134
- "id": "octopus",
135
- "name": "Octopus",
166
+ "llama3": {
167
+ "afterShot": "<|eot_id|>\n\n",
168
+ "assistant": "<|start_header_id|>assistant<|end_header_id|>",
169
+ "id": "llama3",
170
+ "name": "Llama 3",
171
+ "stop": [
172
+ "<|eot_id|>",
173
+ "<|end_of_text|>"
174
+ ],
136
175
  "system": {
137
- "schema": "<|system|>{system}<|end|>",
138
- "message": "You are a router. Below is the query from the users, please call the correct function and generate the parameters to call the function."
176
+ "schema": "<|start_header_id|>system<|end_header_id|>\n\n{system}<|eot_id|>"
139
177
  },
140
- "user": "<|user|>{prompt}<|end|>",
141
- "assistant": "<|assistant|>",
142
- "afterShot": "\n",
143
- "stop": ["<|end|>"]
178
+ "user": "<|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|>"
144
179
  },
145
180
  "llava": {
146
- "id": "llava",
147
- "name": "Llava",
148
- "user": "USER: {prompt}",
149
181
  "assistant": "ASSISTANT:",
182
+ "id": "llava",
150
183
  "linebreaks": {
151
- "user": 1,
184
+ "user": 1
152
185
  },
186
+ "name": "Llava",
187
+ "user": "USER: {prompt}"
153
188
  },
154
- "synthia-cot": {
155
- "id": "synthia-cot",
156
- "name": "Synthia CoT",
157
- "system": {
158
- "schema": "SYSTEM: {system}",
159
- "message": "Elaborate on the topic using a Tree of Thoughts and backtrack when necessary to construct a clear, cohesive Chain of Thought reasoning. Always answer without hesitation."
160
- },
161
- "user": "USER: {prompt}",
162
- "assistant": "ASSISTANT:",
189
+ "minichat": {
190
+ "afterShot": "\n",
191
+ "assistant": "[|Assistant|]",
192
+ "id": "minichat",
193
+ "name": "Minichat",
194
+ "prefix": "<s> ",
195
+ "stop": [
196
+ "</s>",
197
+ "[|User|]"
198
+ ],
199
+ "user": "[|User|] {prompt} </s>"
200
+ },
201
+ "mistral": {
202
+ "afterShot": "\n",
203
+ "assistant": " [/INST]",
204
+ "id": "mistral",
205
+ "name": "Mistral",
206
+ "stop": [
207
+ "</s>"
208
+ ],
209
+ "user": "[INST] {prompt}"
210
+ },
211
+ "nemotron": {
212
+ "afterShot": "\n\n",
213
+ "assistant": "<extra_id_1>Assistant",
214
+ "id": "nemotron",
163
215
  "linebreaks": {
164
- "system": 1,
165
- "user": 1,
216
+ "system": 2,
217
+ "user": 1
166
218
  },
167
- },
168
- "wizardlm": {
169
- "id": "wizardlm",
170
- "name": "WizardLM",
219
+ "name": "Nemotron",
171
220
  "system": {
172
- "schema": "{system}",
173
- "message": "You are a helpful AI assistant."
221
+ "schema": "<extra_id_0>System\n{system}"
174
222
  },
175
- "user": "USER: {prompt}",
176
- "assistant": "ASSISTANT:",
177
- "linebreaks": {
178
- "user": 1
223
+ "user": "<extra_id_1>User\n{prompt}"
224
+ },
225
+ "none": {
226
+ "assistant": "",
227
+ "id": "none",
228
+ "name": "No template",
229
+ "user": "{prompt}"
230
+ },
231
+ "octopus": {
232
+ "afterShot": "\n",
233
+ "assistant": "<|assistant|>",
234
+ "id": "octopus",
235
+ "name": "Octopus",
236
+ "stop": [
237
+ "<|end|>"
238
+ ],
239
+ "system": {
240
+ "message": "You are a router. Below is the query from the users, please call the correct function and generate the parameters to call the function.",
241
+ "schema": "<|system|>{system}<|end|>"
179
242
  },
243
+ "user": "<|user|>{prompt}<|end|>"
180
244
  },
181
245
  "openchat": {
246
+ "assistant": "GPT4 Assistant:",
182
247
  "id": "openchat",
183
248
  "name": "OpenChat",
184
- "user": "GPT4 User: {prompt}<|end_of_turn|>",
185
- "assistant": "GPT4 Assistant:",
186
- "stop": ["<|end_of_turn|>"]
249
+ "stop": [
250
+ "<|end_of_turn|>"
251
+ ],
252
+ "user": "GPT4 User: {prompt}<|end_of_turn|>"
187
253
  },
188
254
  "openchat-correct": {
255
+ "assistant": "GPT4 Correct Assistant:",
189
256
  "id": "openchat-correct",
190
257
  "name": "OpenChat correct",
191
- "user": "GPT4 Correct User: {prompt}<|end_of_turn|>",
192
- "assistant": "GPT4 Correct Assistant:",
193
- "stop": ["<|end_of_turn|>"]
258
+ "stop": [
259
+ "<|end_of_turn|>"
260
+ ],
261
+ "user": "GPT4 Correct User: {prompt}<|end_of_turn|>"
194
262
  },
195
- "human_response": {
196
- "id": "human_response",
197
- "name": "Human response",
198
- "user": "### HUMAN:\n{prompt}",
199
- "assistant": "### RESPONSE:",
263
+ "opencodeinterpreter": {
264
+ "assistant": "<|Assistant|>",
265
+ "id": "opencodeinterpreter",
200
266
  "linebreaks": {
201
- "user": 2,
202
- "assistant": 1
267
+ "user": 2
203
268
  },
269
+ "name": "Open code interpreter",
270
+ "stop": [
271
+ "<|EOT|>",
272
+ "<|User|>"
273
+ ],
274
+ "user": "<|User|>\n{prompt}"
204
275
  },
205
- "minichat": {
206
- "id": "minichat",
207
- "name": "Minichat",
208
- "user": "[|User|] {prompt} </s>",
209
- "assistant": "[|Assistant|]",
210
- "stop": ["</s>", "[|User|]"],
211
- "afterShot": "\n",
212
- "prefix": "<s> "
276
+ "orca": {
277
+ "assistant": "### Response:",
278
+ "id": "orca",
279
+ "linebreaks": {
280
+ "system": 2,
281
+ "user": 2
282
+ },
283
+ "name": "Orca",
284
+ "system": {
285
+ "message": "You are an AI assistant that follows instruction extremely well. Help as much as you can.",
286
+ "schema": "### System:\n{system}"
287
+ },
288
+ "user": "### User:\n{prompt}"
213
289
  },
214
290
  "phi": {
215
- "id": "phi",
216
- "name": "Phi",
217
- "user": "Instruct: {prompt}",
218
291
  "assistant": "Output:",
292
+ "id": "phi",
219
293
  "linebreaks": {
220
294
  "user": 1
221
295
  },
222
- "stop": ["</s>", "Instruct:"]
296
+ "name": "Phi",
297
+ "stop": [
298
+ "</s>",
299
+ "Instruct:"
300
+ ],
301
+ "user": "Instruct: {prompt}"
223
302
  },
224
303
  "phi3": {
304
+ "afterShot": "<|end|>\n",
305
+ "assistant": "<|assistant|>",
225
306
  "id": "phi3",
226
307
  "name": "Phi 3",
227
- "user": "<|user|> {prompt}<|end|>",
228
- "assistant": "<|assistant|>",
308
+ "stop": [
309
+ "<|end|>",
310
+ "<|user|>"
311
+ ],
229
312
  "system": {
230
- "schema": "<|system|> {system}<|end|>",
313
+ "schema": "<|system|> {system}<|end|>"
231
314
  },
232
- "afterShot": "<|end|>\n",
233
- "stop": ["<|end|>", "<|user|>"]
315
+ "user": "<|user|> {prompt}<|end|>"
234
316
  },
235
- "deepseek": {
236
- "id": "deepseek",
237
- "name": "Deepseek",
317
+ "phi4": {
318
+ "afterShot": "<|im_end|>\n",
319
+ "assistant": "<|im_start|>assistant<|im_sep|>",
320
+ "id": "phi4",
321
+ "name": "Phi 4",
322
+ "stop": [
323
+ "<|im_end|>",
324
+ "<|im_sep|>"
325
+ ],
238
326
  "system": {
239
- "schema": "{system}",
240
- "message": "You are an AI programming assistant, utilizing the DeepSeek Coder model, developed by DeepSeek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer."
327
+ "schema": "<|im_start|>system<|im_sep|>{system}<|im_end|>"
241
328
  },
242
- "afterShot": "\n",
243
- "user": "### Instruction:\n{prompt}",
244
- "assistant": "### Response:",
329
+ "user": "<|im_start|>user<|im_sep|>{prompt}<|im_end|>"
330
+ },
331
+ "synthia-cot": {
332
+ "assistant": "ASSISTANT:",
333
+ "id": "synthia-cot",
245
334
  "linebreaks": {
246
- "user": 1,
247
335
  "system": 1,
336
+ "user": 1
248
337
  },
249
- "stop": ["<|EOT|>", "### Instruction:"]
250
- },
251
- "deepseek2": {
252
- "id": "deepseek2",
253
- "name": "Deepseek v2",
338
+ "name": "Synthia CoT",
254
339
  "system": {
255
- "schema": "<|begin▁of▁sentence|>{system}",
340
+ "message": "Elaborate on the topic using a Tree of Thoughts and backtrack when necessary to construct a clear, cohesive Chain of Thought reasoning. Always answer without hesitation.",
341
+ "schema": "SYSTEM: {system}"
256
342
  },
257
- "user": "User: {prompt}",
258
- "assistant": "Assistant:",
343
+ "user": "USER: {prompt}"
344
+ },
345
+ "vicuna": {
346
+ "assistant": "### ASSISTANT:",
347
+ "id": "vicuna",
259
348
  "linebreaks": {
260
- "user": 2,
261
- "system": 2,
349
+ "user": 2
262
350
  },
263
- "stop": ["<|end▁of▁sentence|>", "<|tool▁calls▁end|>"]
351
+ "name": "Vicuna",
352
+ "user": "USER: {prompt}"
264
353
  },
265
- "opencodeinterpreter": {
266
- "id": "opencodeinterpreter",
267
- "name": "Open code interpreter",
268
- "user": "<|User|>\n{prompt}",
269
- "assistant": "<|Assistant|>",
354
+ "vicuna_system": {
355
+ "assistant": "### ASSISTANT:",
356
+ "id": "vicuna_system",
270
357
  "linebreaks": {
358
+ "system": 2,
271
359
  "user": 2
272
360
  },
273
- "stop": ["<|EOT|>", "<|User|>"]
274
- },
275
- "cerebrum": {
276
- "id": "cerebrum",
277
- "name": "Cerebrum",
361
+ "name": "Vicuna system",
278
362
  "system": {
279
- "schema": "{system}",
280
- "message": "A chat between a user and a thinking artificial intelligence assistant. The assistant describes its thought process and gives helpful and detailed answers to the user's questions."
281
- },
282
- "user": "User: {prompt}",
283
- "assistant": "Ai:",
284
- "linebreaks": {
285
- "user": 1
363
+ "schema": "SYSTEM: {system}"
286
364
  },
287
- "prefix": "<s>",
288
- "stop": ["</s>"]
365
+ "user": "USER: {prompt}"
289
366
  },
290
- "command-r": {
291
- "id": "command-r",
292
- "name": "Command-R",
293
- "user": "<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{prompt}<|END_OF_TURN_TOKEN|>",
294
- "assistant": "<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>",
295
- "prefix": "<BOS_TOKEN>",
296
- "stop": [
297
- "<|END_OF_TURN_TOKEN|>"
298
- ],
367
+ "wizard_vicuna": {
368
+ "assistant": "### ASSISTANT:",
369
+ "id": "wizard_vicuna",
299
370
  "linebreaks": {
300
- "user": 1,
371
+ "user": 2
301
372
  },
302
- "system": {
303
- "schema": "<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>{system}<|END_OF_TURN_TOKEN|>"
304
- }
305
- },
306
- "llama3": {
307
- "id": "llama3",
308
- "name": "Llama 3",
309
- "user": "<|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|>",
310
- "assistant": "<|start_header_id|>assistant<|end_header_id|>",
373
+ "name": "Wizard Vicuna",
311
374
  "stop": [
312
- "<|eot_id|>",
313
- "<|end_of_text|>"
375
+ "<|endoftext|>"
314
376
  ],
315
- "afterShot": "<|eot_id|>\n\n",
316
- "system": {
317
- "schema": "<|start_header_id|>system<|end_header_id|>\n\n{system}<|eot_id|>"
318
- }
377
+ "user": "### Human:\n{prompt}"
319
378
  },
320
- "codestral": {
321
- "id": "codestral",
322
- "name": "Codestral",
323
- "user": "[INST] {prompt}",
324
- "assistant": " [/INST]",
325
- "stop": ["</s>"],
326
- "afterShot": "\n",
379
+ "wizardlm": {
380
+ "assistant": "ASSISTANT:",
381
+ "id": "wizardlm",
327
382
  "linebreaks": {
328
- "system": 2,
383
+ "user": 1
329
384
  },
385
+ "name": "WizardLM",
330
386
  "system": {
331
- "schema": "<<SYS>>\n{system}\n<</SYS>>",
387
+ "message": "You are a helpful AI assistant.",
388
+ "schema": "{system}"
332
389
  },
390
+ "user": "USER: {prompt}"
333
391
  },
334
- "gemma": {
335
- "id": "gemma",
336
- "name": "Gemma",
337
- "user": "<start_of_turn>user\n{prompt}",
338
- "assistant": "<end_of_turn>\n<start_of_turn>model",
339
- "stop": ["<end_of_turn>"],
340
- "afterShot": "\n"
341
- },
342
- "nemotron": {
343
- "id": "nemotron",
344
- "name": "Nemotron",
345
- "user": "<extra_id_1>User\n{prompt}",
346
- "assistant": "<extra_id_1>Assistant",
392
+ "zephyr": {
393
+ "afterShot": "\n",
394
+ "assistant": "<|assistant|>",
395
+ "id": "zephyr",
347
396
  "linebreaks": {
348
- "system": 2,
349
- "user": 1,
397
+ "assistant": 1,
398
+ "system": 1,
399
+ "user": 1
350
400
  },
401
+ "name": "Zephyr",
402
+ "stop": [
403
+ "<|endoftext|>"
404
+ ],
351
405
  "system": {
352
- "schema": "<extra_id_0>System\n{system}"
406
+ "schema": "<|system|>\n{system}<|endoftext|>"
353
407
  },
354
- "afterShot": "\n\n"
408
+ "user": "<|user|>\n{prompt}<|endoftext|>"
355
409
  }
356
410
  };
357
411
 
@@ -705,13 +759,13 @@ class PromptTemplate {
705
759
  _buildAssistantBlock(msg) {
706
760
  let buf = [];
707
761
  let amsg = this.assistant;
762
+ if (this?.linebreaks?.assistant) {
763
+ amsg += "\n".repeat(this.linebreaks.assistant);
764
+ }
708
765
  if (this._extraAssistant.length > 0) {
709
766
  amsg += this._extraAssistant;
710
767
  }
711
768
  buf.push(amsg);
712
- if (this?.linebreaks?.assistant) {
713
- buf.push("\n".repeat(this.linebreaks.assistant));
714
- }
715
769
  if (msg) {
716
770
  // this is a shot
717
771
  buf.push(msg);
package/dist/main.min.js CHANGED
@@ -1 +1 @@
1
- var $tpl=function(s){"use strict";const e={none:{id:"none",name:"No template",user:"{prompt}",assistant:""},alpaca:{id:"alpaca",name:"Alpaca",system:{schema:"{system}",message:"Below is an instruction that describes a task. Write a response that appropriately completes the request."},user:"### Instruction:\n{prompt}",assistant:"### Response:",linebreaks:{system:2,user:2}},llama:{id:"llama",name:"Llama",system:{schema:"[INST] <<SYS>>\n{system}\n<</SYS>>",message:"You are a helpful, respectful and honest assistant. Always answer as helpfully as possible\n\nIf a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information."},user:"{prompt}",assistant:" [/INST] ",linebreaks:{system:2,user:0},prefix:"<s>",stop:["</s>"]},mistral:{id:"mistral",name:"Mistral",user:"[INST] {prompt}",assistant:" [/INST]",stop:["</s>"],afterShot:"\n"},orca:{id:"orca",name:"Orca",system:{schema:"### System:\n{system}",message:"You are an AI assistant that follows instruction extremely well. Help as much as you can."},user:"### User:\n{prompt}",assistant:"### Response:",linebreaks:{system:2,user:2}},vicuna:{id:"vicuna",name:"Vicuna",user:"USER: {prompt}",assistant:"### ASSISTANT:",linebreaks:{user:2}},vicuna_system:{id:"vicuna_system",name:"Vicuna system",system:{schema:"SYSTEM: {system}"},user:"USER: {prompt}",assistant:"### ASSISTANT:",linebreaks:{system:2,user:2}},wizard_vicuna:{id:"wizard_vicuna",name:"Wizard Vicuna",user:"### Human:\n{prompt}",assistant:"### ASSISTANT:",linebreaks:{user:2},stop:["<|endoftext|>"]},guanaco:{id:"guanaco",name:"Guanaco",user:"### Human: {prompt}",assistant:"### Assistant:",linebreaks:{user:1}},chatml:{id:"chatml",name:"ChatMl",system:{schema:"<|im_start|>system\n{system}\n<|im_end|>"},user:"<|im_start|>user\n{prompt}<|im_end|>",assistant:"<|im_start|>assistant",linebreaks:{system:1,user:1,assistant:1},stop:["<|im_end|>"],afterShot:" <|im_end|>"},zephyr:{id:"zephyr",name:"Zephyr",system:{schema:"<|system|>\n{system}<|endoftext|>"},user:"<|user|>\n{prompt}<|endoftext|>",assistant:"<|assistant|>",linebreaks:{system:1,user:1,assistant:1},afterShot:"\n",stop:["<|endoftext|>"]},octopus:{id:"octopus",name:"Octopus",system:{schema:"<|system|>{system}<|end|>",message:"You are a router. Below is the query from the users, please call the correct function and generate the parameters to call the function."},user:"<|user|>{prompt}<|end|>",assistant:"<|assistant|>",afterShot:"\n",stop:["<|end|>"]},llava:{id:"llava",name:"Llava",user:"USER: {prompt}",assistant:"ASSISTANT:",linebreaks:{user:1}},"synthia-cot":{id:"synthia-cot",name:"Synthia CoT",system:{schema:"SYSTEM: {system}",message:"Elaborate on the topic using a Tree of Thoughts and backtrack when necessary to construct a clear, cohesive Chain of Thought reasoning. Always answer without hesitation."},user:"USER: {prompt}",assistant:"ASSISTANT:",linebreaks:{system:1,user:1}},wizardlm:{id:"wizardlm",name:"WizardLM",system:{schema:"{system}",message:"You are a helpful AI assistant."},user:"USER: {prompt}",assistant:"ASSISTANT:",linebreaks:{user:1}},openchat:{id:"openchat",name:"OpenChat",user:"GPT4 User: {prompt}<|end_of_turn|>",assistant:"GPT4 Assistant:",stop:["<|end_of_turn|>"]},"openchat-correct":{id:"openchat-correct",name:"OpenChat correct",user:"GPT4 Correct User: {prompt}<|end_of_turn|>",assistant:"GPT4 Correct Assistant:",stop:["<|end_of_turn|>"]},human_response:{id:"human_response",name:"Human response",user:"### HUMAN:\n{prompt}",assistant:"### RESPONSE:",linebreaks:{user:2,assistant:1}},minichat:{id:"minichat",name:"Minichat",user:"[|User|] {prompt} </s>",assistant:"[|Assistant|]",stop:["</s>","[|User|]"],afterShot:"\n",prefix:"<s> "},phi:{id:"phi",name:"Phi",user:"Instruct: {prompt}",assistant:"Output:",linebreaks:{user:1},stop:["</s>","Instruct:"]},phi3:{id:"phi3",name:"Phi 3",user:"<|user|> {prompt}<|end|>",assistant:"<|assistant|>",system:{schema:"<|system|> {system}<|end|>"},afterShot:"<|end|>\n",stop:["<|end|>","<|user|>"]},deepseek:{id:"deepseek",name:"Deepseek",system:{schema:"{system}",message:"You are an AI programming assistant, utilizing the DeepSeek Coder model, developed by DeepSeek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer."},afterShot:"\n",user:"### Instruction:\n{prompt}",assistant:"### Response:",linebreaks:{user:1,system:1},stop:["<|EOT|>","### Instruction:"]},deepseek2:{id:"deepseek2",name:"Deepseek v2",system:{schema:"<|begin▁of▁sentence|>{system}"},user:"User: {prompt}",assistant:"Assistant:",linebreaks:{user:2,system:2},stop:["<|end▁of▁sentence|>","<|tool▁calls▁end|>"]},opencodeinterpreter:{id:"opencodeinterpreter",name:"Open code interpreter",user:"<|User|>\n{prompt}",assistant:"<|Assistant|>",linebreaks:{user:2},stop:["<|EOT|>","<|User|>"]},cerebrum:{id:"cerebrum",name:"Cerebrum",system:{schema:"{system}",message:"A chat between a user and a thinking artificial intelligence assistant. The assistant describes its thought process and gives helpful and detailed answers to the user's questions."},user:"User: {prompt}",assistant:"Ai:",linebreaks:{user:1},prefix:"<s>",stop:["</s>"]},"command-r":{id:"command-r",name:"Command-R",user:"<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{prompt}<|END_OF_TURN_TOKEN|>",assistant:"<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>",prefix:"<BOS_TOKEN>",stop:["<|END_OF_TURN_TOKEN|>"],linebreaks:{user:1},system:{schema:"<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>{system}<|END_OF_TURN_TOKEN|>"}},llama3:{id:"llama3",name:"Llama 3",user:"<|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|>",assistant:"<|start_header_id|>assistant<|end_header_id|>",stop:["<|eot_id|>","<|end_of_text|>"],afterShot:"<|eot_id|>\n\n",system:{schema:"<|start_header_id|>system<|end_header_id|>\n\n{system}<|eot_id|>"}},codestral:{id:"codestral",name:"Codestral",user:"[INST] {prompt}",assistant:" [/INST]",stop:["</s>"],afterShot:"\n",linebreaks:{system:2},system:{schema:"<<SYS>>\n{system}\n<</SYS>>"}},gemma:{id:"gemma",name:"Gemma",user:"<start_of_turn>user\n{prompt}",assistant:"<end_of_turn>\n<start_of_turn>model",stop:["<end_of_turn>"],afterShot:"\n"},nemotron:{id:"nemotron",name:"Nemotron",user:"<extra_id_1>User\n{prompt}",assistant:"<extra_id_1>Assistant",linebreaks:{system:2,user:1},system:{schema:"<extra_id_0>System\n{system}"},afterShot:"\n\n"}};class t{id;name;user;assistant;history=[];system;shots;stop;linebreaks;afterShot;prefix;_extraSystem="";_extraAssistant="";_replacePrompt="";_replaceSystem="";constructor(s){let e;e="string"==typeof s?this._load(s):s,this.id=e.id,this.name=e.name,this.user=e.user,this.assistant=e.assistant,this.system=e.system,this.shots=e.shots,this.stop=e.stop,this.linebreaks=e.linebreaks,this.afterShot=e.afterShot,this.prefix=e.prefix}cloneTo(s,e=!0){const a=new t(s);return e&&this?.shots&&this.shots.forEach((s=>{a.addShot(s.user,s.assistant)})),this._extraSystem.length>0&&a.afterSystem(this._extraSystem),this._replaceSystem.length>0&&a.replaceSystem(this._replaceSystem),this._extraAssistant.length>0&&a.afterAssistant(this._extraAssistant),this._replacePrompt.length>0&&a.replacePrompt(this._replacePrompt),a}toJson(){const s={id:this.id,name:this.name,user:this.user,assistant:this.assistant};return this?.prefix&&(s.prefix=this.prefix),this?.system&&(s.system=this.system),this?.shots&&(s.shots=this.shots),this?.afterShot&&(s.afterShot=this.afterShot),this?.stop&&(s.stop=this.stop),this?.linebreaks&&(s.linebreaks=this.linebreaks),s}replaceSystem(s){return this.system?(this._replaceSystem=s,this):this}afterSystem(s){return this.system?(this._extraSystem=s,this):this}afterAssistant(s){return this._extraAssistant=s,this}replacePrompt(s){return this._replacePrompt=s,this}addShot(s,e){this?.shots||(this.shots=[]);let t=e;return this.shots.push({user:s,assistant:t}),this}addShots(s){return s.forEach((s=>this.addShot(s.user,s.assistant))),this}renderShot(s){const e=[];e.push(this._buildUserBlock(s.user));let t=s.assistant;return this.afterShot&&(t+=this.afterShot),e.push(this._buildAssistantBlock(t)),e.join("")}render(s=!1){const e=new Array;this.prefix&&e.push(this.prefix);const t=this._buildSystemBlock(s);if(t.length>0&&(e.push(t),this?.linebreaks?.system&&e.push("\n".repeat(this.linebreaks.system))),this?.shots)for(const s of this.shots)e.push(this.renderShot(s));for(const s of this.history)e.push(this.renderShot(s));return e.push(this._buildUserBlock()),e.push(this._buildAssistantBlock()),e.join("")}prompt(s){return this.render().replace("{prompt}",s)}pushToHistory(s){return this.history.push(s),this}_buildSystemBlock(s){let e="";return this?.system?(this._replaceSystem&&(this.system.message=this._replaceSystem),this.system?.message?(e=this.system.schema.replace("{system}",this.system.message),this._extraSystem&&(e+=this._extraSystem)):s||(e=this.system.schema),e):""}_buildUserBlock(s){let e=[],t=this.user;return this._replacePrompt.length>0&&(t=t.replace("{prompt}",this._replacePrompt)),e.push(t),this?.linebreaks?.user&&e.push("\n".repeat(this.linebreaks.user)),s&&(e[0]=this.user.replace("{prompt}",s)),e.join("")}_buildAssistantBlock(s){let e=[],t=this.assistant;return this._extraAssistant.length>0&&(t+=this._extraAssistant),e.push(t),this?.linebreaks?.assistant&&e.push("\n".repeat(this.linebreaks.assistant)),s&&e.push(s),e.join("")}_load(s){try{if(s in e)return e[s];throw new Error(`Template ${s} not found`)}catch(e){throw new Error(`Error loading template ${s}: ${e}`)}}}return s.PromptTemplate=t,s.templates=e,s}({});
1
+ var $tpl=function(s){"use strict";const e={alpaca:{assistant:"### Response:",id:"alpaca",linebreaks:{system:2,user:2},name:"Alpaca",system:{message:"Below is an instruction that describes a task. Write a response that appropriately completes the request.",schema:"{system}"},user:"### Instruction:\n{prompt}"},cerebrum:{assistant:"Ai:",id:"cerebrum",linebreaks:{user:1},name:"Cerebrum",prefix:"<s>",stop:["</s>"],system:{message:"A chat between a user and a thinking artificial intelligence assistant. The assistant describes its thought process and gives helpful and detailed answers to the user's questions.",schema:"{system}"},user:"User: {prompt}"},chatml:{afterShot:" <|im_end|>\n",assistant:"<|im_start|>assistant",id:"chatml",linebreaks:{assistant:1,system:1,user:1},name:"ChatMl",stop:["<|im_end|>"],system:{schema:"<|im_start|>system\n{system}<|im_end|>"},user:"<|im_start|>user\n{prompt}<|im_end|>"},codestral:{afterShot:"\n",assistant:" [/INST]",id:"codestral",linebreaks:{system:2},name:"Codestral",stop:["</s>"],system:{schema:"<<SYS>>\n{system}\n<</SYS>>"},user:"[INST] {prompt}"},"command-r":{assistant:"<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>",id:"command-r",linebreaks:{user:1},name:"Command-R",prefix:"<BOS_TOKEN>",stop:["<|END_OF_TURN_TOKEN|>"],system:{schema:"<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>{system}<|END_OF_TURN_TOKEN|>"},user:"<|START_OF_TURN_TOKEN|><|USER_TOKEN|>{prompt}<|END_OF_TURN_TOKEN|>"},deepseek:{afterShot:"\n",assistant:"### Response:",id:"deepseek",linebreaks:{system:1,user:1},name:"Deepseek",stop:["<|EOT|>","### Instruction:"],system:{message:"You are an AI programming assistant, utilizing the DeepSeek Coder model, developed by DeepSeek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer.",schema:"{system}"},user:"### Instruction:\n{prompt}"},deepseek2:{assistant:"Assistant:",id:"deepseek2",linebreaks:{system:2,user:2},name:"Deepseek v2",stop:["<|end▁of▁sentence|>","<|tool▁calls▁end|>"],system:{schema:"<|begin▁of▁sentence|>{system}"},user:"User: {prompt}"},gemma:{afterShot:"\n",assistant:"<end_of_turn>\n<start_of_turn>model",id:"gemma",name:"Gemma",stop:["<end_of_turn>"],user:"<start_of_turn>user\n{prompt}"},guanaco:{assistant:"### Assistant:",id:"guanaco",linebreaks:{user:1},name:"Guanaco",user:"### Human: {prompt}"},human_response:{assistant:"### RESPONSE:",id:"human_response",linebreaks:{assistant:1,user:2},name:"Human response",user:"### HUMAN:\n{prompt}"},llama:{assistant:" [/INST] ",id:"llama",linebreaks:{system:2,user:0},name:"Llama",prefix:"<s>",stop:["</s>"],system:{message:"You are a helpful, respectful and honest assistant. Always answer as helpfully as possible\n\nIf a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.",schema:"[INST] <<SYS>>\n{system}\n<</SYS>>"},user:"{prompt}"},llama3:{afterShot:"<|eot_id|>\n\n",assistant:"<|start_header_id|>assistant<|end_header_id|>",id:"llama3",name:"Llama 3",stop:["<|eot_id|>","<|end_of_text|>"],system:{schema:"<|start_header_id|>system<|end_header_id|>\n\n{system}<|eot_id|>"},user:"<|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|>"},llava:{assistant:"ASSISTANT:",id:"llava",linebreaks:{user:1},name:"Llava",user:"USER: {prompt}"},minichat:{afterShot:"\n",assistant:"[|Assistant|]",id:"minichat",name:"Minichat",prefix:"<s> ",stop:["</s>","[|User|]"],user:"[|User|] {prompt} </s>"},mistral:{afterShot:"\n",assistant:" [/INST]",id:"mistral",name:"Mistral",stop:["</s>"],user:"[INST] {prompt}"},nemotron:{afterShot:"\n\n",assistant:"<extra_id_1>Assistant",id:"nemotron",linebreaks:{system:2,user:1},name:"Nemotron",system:{schema:"<extra_id_0>System\n{system}"},user:"<extra_id_1>User\n{prompt}"},none:{assistant:"",id:"none",name:"No template",user:"{prompt}"},octopus:{afterShot:"\n",assistant:"<|assistant|>",id:"octopus",name:"Octopus",stop:["<|end|>"],system:{message:"You are a router. Below is the query from the users, please call the correct function and generate the parameters to call the function.",schema:"<|system|>{system}<|end|>"},user:"<|user|>{prompt}<|end|>"},openchat:{assistant:"GPT4 Assistant:",id:"openchat",name:"OpenChat",stop:["<|end_of_turn|>"],user:"GPT4 User: {prompt}<|end_of_turn|>"},"openchat-correct":{assistant:"GPT4 Correct Assistant:",id:"openchat-correct",name:"OpenChat correct",stop:["<|end_of_turn|>"],user:"GPT4 Correct User: {prompt}<|end_of_turn|>"},opencodeinterpreter:{assistant:"<|Assistant|>",id:"opencodeinterpreter",linebreaks:{user:2},name:"Open code interpreter",stop:["<|EOT|>","<|User|>"],user:"<|User|>\n{prompt}"},orca:{assistant:"### Response:",id:"orca",linebreaks:{system:2,user:2},name:"Orca",system:{message:"You are an AI assistant that follows instruction extremely well. Help as much as you can.",schema:"### System:\n{system}"},user:"### User:\n{prompt}"},phi:{assistant:"Output:",id:"phi",linebreaks:{user:1},name:"Phi",stop:["</s>","Instruct:"],user:"Instruct: {prompt}"},phi3:{afterShot:"<|end|>\n",assistant:"<|assistant|>",id:"phi3",name:"Phi 3",stop:["<|end|>","<|user|>"],system:{schema:"<|system|> {system}<|end|>"},user:"<|user|> {prompt}<|end|>"},phi4:{afterShot:"<|im_end|>\n",assistant:"<|im_start|>assistant<|im_sep|>",id:"phi4",name:"Phi 4",stop:["<|im_end|>","<|im_sep|>"],system:{schema:"<|im_start|>system<|im_sep|>{system}<|im_end|>"},user:"<|im_start|>user<|im_sep|>{prompt}<|im_end|>"},"synthia-cot":{assistant:"ASSISTANT:",id:"synthia-cot",linebreaks:{system:1,user:1},name:"Synthia CoT",system:{message:"Elaborate on the topic using a Tree of Thoughts and backtrack when necessary to construct a clear, cohesive Chain of Thought reasoning. Always answer without hesitation.",schema:"SYSTEM: {system}"},user:"USER: {prompt}"},vicuna:{assistant:"### ASSISTANT:",id:"vicuna",linebreaks:{user:2},name:"Vicuna",user:"USER: {prompt}"},vicuna_system:{assistant:"### ASSISTANT:",id:"vicuna_system",linebreaks:{system:2,user:2},name:"Vicuna system",system:{schema:"SYSTEM: {system}"},user:"USER: {prompt}"},wizard_vicuna:{assistant:"### ASSISTANT:",id:"wizard_vicuna",linebreaks:{user:2},name:"Wizard Vicuna",stop:["<|endoftext|>"],user:"### Human:\n{prompt}"},wizardlm:{assistant:"ASSISTANT:",id:"wizardlm",linebreaks:{user:1},name:"WizardLM",system:{message:"You are a helpful AI assistant.",schema:"{system}"},user:"USER: {prompt}"},zephyr:{afterShot:"\n",assistant:"<|assistant|>",id:"zephyr",linebreaks:{assistant:1,system:1,user:1},name:"Zephyr",stop:["<|endoftext|>"],system:{schema:"<|system|>\n{system}<|endoftext|>"},user:"<|user|>\n{prompt}<|endoftext|>"}};class t{id;name;user;assistant;history=[];system;shots;stop;linebreaks;afterShot;prefix;_extraSystem="";_extraAssistant="";_replacePrompt="";_replaceSystem="";constructor(s){let e;e="string"==typeof s?this._load(s):s,this.id=e.id,this.name=e.name,this.user=e.user,this.assistant=e.assistant,this.system=e.system,this.shots=e.shots,this.stop=e.stop,this.linebreaks=e.linebreaks,this.afterShot=e.afterShot,this.prefix=e.prefix}cloneTo(s,e=!0){const a=new t(s);return e&&this?.shots&&this.shots.forEach((s=>{a.addShot(s.user,s.assistant)})),this._extraSystem.length>0&&a.afterSystem(this._extraSystem),this._replaceSystem.length>0&&a.replaceSystem(this._replaceSystem),this._extraAssistant.length>0&&a.afterAssistant(this._extraAssistant),this._replacePrompt.length>0&&a.replacePrompt(this._replacePrompt),a}toJson(){const s={id:this.id,name:this.name,user:this.user,assistant:this.assistant};return this?.prefix&&(s.prefix=this.prefix),this?.system&&(s.system=this.system),this?.shots&&(s.shots=this.shots),this?.afterShot&&(s.afterShot=this.afterShot),this?.stop&&(s.stop=this.stop),this?.linebreaks&&(s.linebreaks=this.linebreaks),s}replaceSystem(s){return this.system?(this._replaceSystem=s,this):this}afterSystem(s){return this.system?(this._extraSystem=s,this):this}afterAssistant(s){return this._extraAssistant=s,this}replacePrompt(s){return this._replacePrompt=s,this}addShot(s,e){this?.shots||(this.shots=[]);let t=e;return this.shots.push({user:s,assistant:t}),this}addShots(s){return s.forEach((s=>this.addShot(s.user,s.assistant))),this}renderShot(s){const e=[];e.push(this._buildUserBlock(s.user));let t=s.assistant;return this.afterShot&&(t+=this.afterShot),e.push(this._buildAssistantBlock(t)),e.join("")}render(s=!1){const e=new Array;this.prefix&&e.push(this.prefix);const t=this._buildSystemBlock(s);if(t.length>0&&(e.push(t),this?.linebreaks?.system&&e.push("\n".repeat(this.linebreaks.system))),this?.shots)for(const s of this.shots)e.push(this.renderShot(s));for(const s of this.history)e.push(this.renderShot(s));return e.push(this._buildUserBlock()),e.push(this._buildAssistantBlock()),e.join("")}prompt(s){return this.render().replace("{prompt}",s)}pushToHistory(s){return this.history.push(s),this}_buildSystemBlock(s){let e="";return this?.system?(this._replaceSystem&&(this.system.message=this._replaceSystem),this.system?.message?(e=this.system.schema.replace("{system}",this.system.message),this._extraSystem&&(e+=this._extraSystem)):s||(e=this.system.schema),e):""}_buildUserBlock(s){let e=[],t=this.user;return this._replacePrompt.length>0&&(t=t.replace("{prompt}",this._replacePrompt)),e.push(t),this?.linebreaks?.user&&e.push("\n".repeat(this.linebreaks.user)),s&&(e[0]=this.user.replace("{prompt}",s)),e.join("")}_buildAssistantBlock(s){let e=[],t=this.assistant;return this?.linebreaks?.assistant&&(t+="\n".repeat(this.linebreaks.assistant)),this._extraAssistant.length>0&&(t+=this._extraAssistant),e.push(t),s&&e.push(s),e.join("")}_load(s){try{if(s in e)return e[s];throw new Error(`Template ${s} not found`)}catch(e){throw new Error(`Error loading template ${s}: ${e}`)}}}return s.PromptTemplate=t,s.templates=e,s}({});
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "modprompt",
3
- "version": "0.9.2",
3
+ "version": "0.9.3",
4
4
  "description": "Prompt templates for language models",
5
5
  "license": "MIT",
6
6
  "scripts": {