@genai-fi/nanogpt 0.15.8 → 0.15.9
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/Generator.d.ts +17 -1
- package/dist/TeachableLLM.d.ts +2 -2
- package/dist/main.d.ts +1 -1
- package/package.json +1 -1
package/dist/Generator.d.ts
CHANGED
|
@@ -19,11 +19,27 @@ export interface IGenerateOptions extends GenerateOptions {
|
|
|
19
19
|
nonConversational?: boolean;
|
|
20
20
|
continuation?: boolean;
|
|
21
21
|
}
|
|
22
|
+
export interface IGenerator extends EE<'start' | 'stop' | 'tokens'> {
|
|
23
|
+
generate(prompt: Conversation[], options?: IGenerateOptions): Promise<Conversation[]>;
|
|
24
|
+
generate(options?: IGenerateOptions): Promise<Conversation[]>;
|
|
25
|
+
step(prompt: Conversation[], options?: IGenerateOptions): Promise<Conversation[]>;
|
|
26
|
+
step(options?: IGenerateOptions): Promise<Conversation[]>;
|
|
27
|
+
stop(): void;
|
|
28
|
+
getConversation(): Conversation[];
|
|
29
|
+
getAttentionData(): number[][][][][];
|
|
30
|
+
getProbabilitiesData(): number[][][];
|
|
31
|
+
getEmbeddingsData(): {
|
|
32
|
+
name: string;
|
|
33
|
+
tensor: number[][];
|
|
34
|
+
}[][];
|
|
35
|
+
getTokens(): number[];
|
|
36
|
+
getLastLoss(): number | null;
|
|
37
|
+
}
|
|
22
38
|
/**
|
|
23
39
|
* Text generator using a NanoGPT model and a tokeniser.
|
|
24
40
|
* This uses the forward method of the model to generate text token by token, including options for temperature, top-k, and top-p sampling.
|
|
25
41
|
*/
|
|
26
|
-
export default class Generator extends EE<'start' | 'stop' | 'tokens'> {
|
|
42
|
+
export default class Generator extends EE<'start' | 'stop' | 'tokens'> implements IGenerator {
|
|
27
43
|
private readonly model;
|
|
28
44
|
private readonly tokeniser;
|
|
29
45
|
private active;
|
package/dist/TeachableLLM.d.ts
CHANGED
|
@@ -2,7 +2,7 @@ import { GPTConfig } from './models/config';
|
|
|
2
2
|
import { Conversation, ITokeniser } from './tokeniser/type';
|
|
3
3
|
import { SaveOptions } from './loader/save';
|
|
4
4
|
import { LoadModelOptions } from './loader/load';
|
|
5
|
-
import {
|
|
5
|
+
import { IGenerateOptions, IGenerator } from './Generator';
|
|
6
6
|
import { default as Trainer, TrainingType } from './Trainer';
|
|
7
7
|
import { default as MemoryProfiler } from './utilities/profile';
|
|
8
8
|
import { default as Model, ModelForwardAttributes } from './models/model';
|
|
@@ -50,7 +50,7 @@ export default class TeachableLLM {
|
|
|
50
50
|
trainer(trainingType?: TrainingType, options?: TrainingOptions): Trainer;
|
|
51
51
|
train(text: Task[], options?: TrainingOptions, trainingType?: TrainingType): Promise<void>;
|
|
52
52
|
trainTokeniser(text: string[]): Promise<number>;
|
|
53
|
-
generator():
|
|
53
|
+
generator(): IGenerator;
|
|
54
54
|
generateText(prompt: Conversation[], options?: IGenerateOptions): Promise<Conversation[]>;
|
|
55
55
|
generateText(options?: IGenerateOptions): Promise<Conversation[]>;
|
|
56
56
|
dispose(): void;
|
package/dist/main.d.ts
CHANGED
|
@@ -13,7 +13,7 @@ export { default as CharTokeniser } from './tokeniser/CharTokeniser';
|
|
|
13
13
|
export { default as BPETokeniser } from './tokeniser/bpe';
|
|
14
14
|
export { default as waitForModel } from './utilities/waitForModel';
|
|
15
15
|
export { default as loadTextData } from './data/textLoader';
|
|
16
|
-
export { default as Generator } from './Generator';
|
|
16
|
+
export { default as Generator, type IGenerator } from './Generator';
|
|
17
17
|
export { default as Evaluator } from './training/Evaluator';
|
|
18
18
|
export { default as Trainer } from './Trainer';
|
|
19
19
|
export type { IGenerateOptions } from './Generator';
|