@push.rocks/smartai 0.2.0 → 0.3.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3,7 +3,7 @@
3
3
  */
4
4
  export const commitinfo = {
5
5
  name: '@push.rocks/smartai',
6
- version: '0.2.0',
6
+ version: '0.3.0',
7
7
  description: 'A TypeScript library for integrating and interacting with multiple AI models, offering capabilities for chat and potentially audio responses.'
8
8
  };
9
9
  //# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiMDBfY29tbWl0aW5mb19kYXRhLmpzIiwic291cmNlUm9vdCI6IiIsInNvdXJjZXMiOlsiLi4vdHMvMDBfY29tbWl0aW5mb19kYXRhLnRzIl0sIm5hbWVzIjpbXSwibWFwcGluZ3MiOiJBQUFBOztHQUVHO0FBQ0gsTUFBTSxDQUFDLE1BQU0sVUFBVSxHQUFHO0lBQ3hCLElBQUksRUFBRSxxQkFBcUI7SUFDM0IsT0FBTyxFQUFFLE9BQU87SUFDaEIsV0FBVyxFQUFFLCtJQUErSTtDQUM3SixDQUFBIn0=
@@ -0,0 +1,43 @@
1
+ import * as plugins from './plugins.js';
2
+ import { MultiModalModel } from './abstract.classes.multimodal.js';
3
+ export interface IXAIProviderOptions {
4
+ xaiToken: string;
5
+ }
6
+ export declare class XAIProvider extends MultiModalModel {
7
+ private options;
8
+ openAiApiClient: plugins.openai.default;
9
+ smartpdfInstance: plugins.smartpdf.SmartPdf;
10
+ constructor(optionsArg: IXAIProviderOptions);
11
+ start(): Promise<void>;
12
+ stop(): Promise<void>;
13
+ chatStream(input: ReadableStream<Uint8Array>): Promise<ReadableStream<string>>;
14
+ chat(optionsArg: {
15
+ systemMessage: string;
16
+ userMessage: string;
17
+ messageHistory: {
18
+ role: string;
19
+ content: string;
20
+ }[];
21
+ }): Promise<{
22
+ role: 'assistant';
23
+ message: string;
24
+ }>;
25
+ audio(optionsArg: {
26
+ message: string;
27
+ }): Promise<NodeJS.ReadableStream>;
28
+ vision(optionsArg: {
29
+ image: Buffer;
30
+ prompt: string;
31
+ }): Promise<string>;
32
+ document(optionsArg: {
33
+ systemMessage: string;
34
+ userMessage: string;
35
+ pdfDocuments: Uint8Array[];
36
+ messageHistory: {
37
+ role: string;
38
+ content: string;
39
+ }[];
40
+ }): Promise<{
41
+ message: any;
42
+ }>;
43
+ }
@@ -0,0 +1,141 @@
1
+ import * as plugins from './plugins.js';
2
+ import * as paths from './paths.js';
3
+ import { MultiModalModel } from './abstract.classes.multimodal.js';
4
+ export class XAIProvider extends MultiModalModel {
5
+ constructor(optionsArg) {
6
+ super();
7
+ this.options = optionsArg;
8
+ }
9
+ async start() {
10
+ this.openAiApiClient = new plugins.openai.default({
11
+ apiKey: this.options.xaiToken,
12
+ baseURL: 'https://api.x.ai/v1',
13
+ });
14
+ this.smartpdfInstance = new plugins.smartpdf.SmartPdf();
15
+ }
16
+ async stop() { }
17
+ async chatStream(input) {
18
+ // Create a TextDecoder to handle incoming chunks
19
+ const decoder = new TextDecoder();
20
+ let buffer = '';
21
+ let currentMessage = null;
22
+ // Create a TransformStream to process the input
23
+ const transform = new TransformStream({
24
+ async transform(chunk, controller) {
25
+ buffer += decoder.decode(chunk, { stream: true });
26
+ // Try to parse complete JSON messages from the buffer
27
+ while (true) {
28
+ const newlineIndex = buffer.indexOf('\n');
29
+ if (newlineIndex === -1)
30
+ break;
31
+ const line = buffer.slice(0, newlineIndex);
32
+ buffer = buffer.slice(newlineIndex + 1);
33
+ if (line.trim()) {
34
+ try {
35
+ const message = JSON.parse(line);
36
+ currentMessage = {
37
+ role: message.role || 'user',
38
+ content: message.content || '',
39
+ };
40
+ }
41
+ catch (e) {
42
+ console.error('Failed to parse message:', e);
43
+ }
44
+ }
45
+ }
46
+ // If we have a complete message, send it to X.AI
47
+ if (currentMessage) {
48
+ const stream = await this.openAiApiClient.chat.completions.create({
49
+ model: 'grok-2-latest',
50
+ messages: [{ role: currentMessage.role, content: currentMessage.content }],
51
+ stream: true,
52
+ });
53
+ // Process each chunk from X.AI
54
+ for await (const chunk of stream) {
55
+ const content = chunk.choices[0]?.delta?.content;
56
+ if (content) {
57
+ controller.enqueue(content);
58
+ }
59
+ }
60
+ currentMessage = null;
61
+ }
62
+ },
63
+ flush(controller) {
64
+ if (buffer) {
65
+ try {
66
+ const message = JSON.parse(buffer);
67
+ controller.enqueue(message.content || '');
68
+ }
69
+ catch (e) {
70
+ console.error('Failed to parse remaining buffer:', e);
71
+ }
72
+ }
73
+ }
74
+ });
75
+ // Connect the input to our transform stream
76
+ return input.pipeThrough(transform);
77
+ }
78
+ async chat(optionsArg) {
79
+ // Prepare messages array with system message, history, and user message
80
+ const messages = [
81
+ { role: 'system', content: optionsArg.systemMessage },
82
+ ...optionsArg.messageHistory.map(msg => ({
83
+ role: msg.role,
84
+ content: msg.content
85
+ })),
86
+ { role: 'user', content: optionsArg.userMessage }
87
+ ];
88
+ // Call X.AI's chat completion API
89
+ const completion = await this.openAiApiClient.chat.completions.create({
90
+ model: 'grok-2-latest',
91
+ messages: messages,
92
+ stream: false,
93
+ });
94
+ // Return the assistant's response
95
+ return {
96
+ role: 'assistant',
97
+ message: completion.choices[0]?.message?.content || ''
98
+ };
99
+ }
100
+ async audio(optionsArg) {
101
+ throw new Error('Audio generation is not supported by X.AI');
102
+ }
103
+ async vision(optionsArg) {
104
+ throw new Error('Vision tasks are not supported by X.AI');
105
+ }
106
+ async document(optionsArg) {
107
+ // First convert PDF documents to images
108
+ let pdfDocumentImageBytesArray = [];
109
+ for (const pdfDocument of optionsArg.pdfDocuments) {
110
+ const documentImageArray = await this.smartpdfInstance.convertPDFToPngBytes(pdfDocument);
111
+ pdfDocumentImageBytesArray = pdfDocumentImageBytesArray.concat(documentImageArray);
112
+ }
113
+ // Convert images to base64 for inclusion in the message
114
+ const imageBase64Array = pdfDocumentImageBytesArray.map(bytes => Buffer.from(bytes).toString('base64'));
115
+ // Combine document images into the user message
116
+ const enhancedUserMessage = `
117
+ ${optionsArg.userMessage}
118
+
119
+ Document contents (as images):
120
+ ${imageBase64Array.map((img, i) => `Image ${i + 1}: <image data>`).join('\n')}
121
+ `;
122
+ // Use chat completion to analyze the documents
123
+ const messages = [
124
+ { role: 'system', content: optionsArg.systemMessage },
125
+ ...optionsArg.messageHistory.map(msg => ({
126
+ role: msg.role,
127
+ content: msg.content
128
+ })),
129
+ { role: 'user', content: enhancedUserMessage }
130
+ ];
131
+ const completion = await this.openAiApiClient.chat.completions.create({
132
+ model: 'grok-2-latest',
133
+ messages: messages,
134
+ stream: false,
135
+ });
136
+ return {
137
+ message: completion.choices[0]?.message?.content || ''
138
+ };
139
+ }
140
+ }
141
+ //# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoicHJvdmlkZXIueGFpLmpzIiwic291cmNlUm9vdCI6IiIsInNvdXJjZXMiOlsiLi4vdHMvcHJvdmlkZXIueGFpLnRzIl0sIm5hbWVzIjpbXSwibWFwcGluZ3MiOiJBQUFBLE9BQU8sS0FBSyxPQUFPLE1BQU0sY0FBYyxDQUFDO0FBQ3hDLE9BQU8sS0FBSyxLQUFLLE1BQU0sWUFBWSxDQUFDO0FBQ3BDLE9BQU8sRUFBRSxlQUFlLEVBQUUsTUFBTSxrQ0FBa0MsQ0FBQztBQVFuRSxNQUFNLE9BQU8sV0FBWSxTQUFRLGVBQWU7SUFLOUMsWUFBWSxVQUErQjtRQUN6QyxLQUFLLEVBQUUsQ0FBQztRQUNSLElBQUksQ0FBQyxPQUFPLEdBQUcsVUFBVSxDQUFDO0lBQzVCLENBQUM7SUFFTSxLQUFLLENBQUMsS0FBSztRQUNoQixJQUFJLENBQUMsZUFBZSxHQUFHLElBQUksT0FBTyxDQUFDLE1BQU0sQ0FBQyxPQUFPLENBQUM7WUFDaEQsTUFBTSxFQUFFLElBQUksQ0FBQyxPQUFPLENBQUMsUUFBUTtZQUM3QixPQUFPLEVBQUUscUJBQXFCO1NBQy9CLENBQUMsQ0FBQztRQUNILElBQUksQ0FBQyxnQkFBZ0IsR0FBRyxJQUFJLE9BQU8sQ0FBQyxRQUFRLENBQUMsUUFBUSxFQUFFLENBQUM7SUFDMUQsQ0FBQztJQUVNLEtBQUssQ0FBQyxJQUFJLEtBQUksQ0FBQztJQUVmLEtBQUssQ0FBQyxVQUFVLENBQUMsS0FBaUM7UUFDdkQsaURBQWlEO1FBQ2pELE1BQU0sT0FBTyxHQUFHLElBQUksV0FBVyxFQUFFLENBQUM7UUFDbEMsSUFBSSxNQUFNLEdBQUcsRUFBRSxDQUFDO1FBQ2hCLElBQUksY0FBYyxHQUE4QyxJQUFJLENBQUM7UUFFckUsZ0RBQWdEO1FBQ2hELE1BQU0sU0FBUyxHQUFHLElBQUksZUFBZSxDQUFxQjtZQUN4RCxLQUFLLENBQUMsU0FBUyxDQUFDLEtBQUssRUFBRSxVQUFVO2dCQUMvQixNQUFNLElBQUksT0FBTyxDQUFDLE1BQU0sQ0FBQyxLQUFLLEVBQUUsRUFBRSxNQUFNLEVBQUUsSUFBSSxFQUFFLENBQUMsQ0FBQztnQkFFbEQsc0RBQXNEO2dCQUN0RCxPQUFPLElBQUksRUFBRSxDQUFDO29CQUNaLE1BQU0sWUFBWSxHQUFHLE1BQU0sQ0FBQyxPQUFPLENBQUMsSUFBSSxDQUFDLENBQUM7b0JBQzFDLElBQUksWUFBWSxLQUFLLENBQUMsQ0FBQzt3QkFBRSxNQUFNO29CQUUvQixNQUFNLElBQUksR0FBRyxNQUFNLENBQUMsS0FBSyxDQUFDLENBQUMsRUFBRSxZQUFZLENBQUMsQ0FBQztvQkFDM0MsTUFBTSxHQUFHLE1BQU0sQ0FBQyxLQUFLLENBQUMsWUFBWSxHQUFHLENBQUMsQ0FBQyxDQUFDO29CQUV4QyxJQUFJLElBQUksQ0FBQyxJQUFJLEVBQUUsRUFBRSxDQUFDO3dCQUNoQixJQUFJLENBQUM7NEJBQ0gsTUFBTSxPQUFPLEdBQUcsSUFBSSxDQUFDLEtBQUssQ0FBQyxJQUFJLENBQUMsQ0FBQzs0QkFDakMsY0FBYyxHQUFHO2dDQUNmLElBQUksRUFBRSxPQUFPLENBQUMsSUFBSSxJQUFJLE1BQU07Z0NBQzVCLE9BQU8sRUFBRSxPQUFPLENBQUMsT0FBTyxJQUFJLEVBQUU7NkJBQy9CLENBQUM7d0JBQ0osQ0FBQzt3QkFBQyxPQUFPLENBQUMsRUFBRSxDQUFDOzRCQUNYLE9BQU8sQ0FBQyxLQUFLLENBQUMsMEJBQTBCLEVBQUUsQ0FBQyxDQUFDLENBQUM7d0JBQy9DLENBQUM7b0JBQ0gsQ0FBQztnQkFDSCxDQUFDO2dCQUVELGlEQUFpRDtnQkFDakQsSUFBSSxjQUFjLEVBQUUsQ0FBQztvQkFDbkIsTUFBTSxNQUFNLEdBQUcsTUFBTSxJQUFJLENBQUMsZUFBZSxDQUFDLElBQUksQ0FBQyxXQUFXLENBQUMsTUFBTSxDQUFDO3dCQUNoRSxLQUFLLEVBQUUsZUFBZTt3QkFDdEIsUUFBUSxFQUFFLENBQUMsRUFBRSxJQUFJLEVBQUUsY0FBYyxDQUFDLElBQUksRUFBRSxPQUFPLEVBQUUsY0FBYyxDQUFDLE9BQU8sRUFBRSxDQUFDO3dCQUMxRSxNQUFNLEVBQUUsSUFBSTtxQkFDYixDQUFDLENBQUM7b0JBRUgsK0JBQStCO29CQUMvQixJQUFJLEtBQUssRUFBRSxNQUFNLEtBQUssSUFBSSxNQUFNLEVBQUUsQ0FBQzt3QkFDakMsTUFBTSxPQUFPLEdBQUcsS0FBSyxDQUFDLE9BQU8sQ0FBQyxDQUFDLENBQUMsRUFBRSxLQUFLLEVBQUUsT0FBTyxDQUFDO3dCQUNqRCxJQUFJLE9BQU8sRUFBRSxDQUFDOzRCQUNaLFVBQVUsQ0FBQyxPQUFPLENBQUMsT0FBTyxDQUFDLENBQUM7d0JBQzlCLENBQUM7b0JBQ0gsQ0FBQztvQkFFRCxjQUFjLEdBQUcsSUFBSSxDQUFDO2dCQUN4QixDQUFDO1lBQ0gsQ0FBQztZQUVELEtBQUssQ0FBQyxVQUFVO2dCQUNkLElBQUksTUFBTSxFQUFFLENBQUM7b0JBQ1gsSUFBSSxDQUFDO3dCQUNILE1BQU0sT0FBTyxHQUFHLElBQUksQ0FBQyxLQUFLLENBQUMsTUFBTSxDQUFDLENBQUM7d0JBQ25DLFVBQVUsQ0FBQyxPQUFPLENBQUMsT0FBTyxDQUFDLE9BQU8sSUFBSSxFQUFFLENBQUMsQ0FBQztvQkFDNUMsQ0FBQztvQkFBQyxPQUFPLENBQUMsRUFBRSxDQUFDO3dCQUNYLE9BQU8sQ0FBQyxLQUFLLENBQUMsbUNBQW1DLEVBQUUsQ0FBQyxDQUFDLENBQUM7b0JBQ3hELENBQUM7Z0JBQ0gsQ0FBQztZQUNILENBQUM7U0FDRixDQUFDLENBQUM7UUFFSCw0Q0FBNEM7UUFDNUMsT0FBTyxLQUFLLENBQUMsV0FBVyxDQUFDLFNBQVMsQ0FBQyxDQUFDO0lBQ3RDLENBQUM7SUFFTSxLQUFLLENBQUMsSUFBSSxDQUFDLFVBSWpCO1FBQ0Msd0VBQXdFO1FBQ3hFLE1BQU0sUUFBUSxHQUFpQztZQUM3QyxFQUFFLElBQUksRUFBRSxRQUFRLEVBQUUsT0FBTyxFQUFFLFVBQVUsQ0FBQyxhQUFhLEVBQUU7WUFDckQsR0FBRyxVQUFVLENBQUMsY0FBYyxDQUFDLEdBQUcsQ0FBQyxHQUFHLENBQUMsRUFBRSxDQUFDLENBQUM7Z0JBQ3ZDLElBQUksRUFBRSxHQUFHLENBQUMsSUFBdUM7Z0JBQ2pELE9BQU8sRUFBRSxHQUFHLENBQUMsT0FBTzthQUNyQixDQUFDLENBQUM7WUFDSCxFQUFFLElBQUksRUFBRSxNQUFNLEVBQUUsT0FBTyxFQUFFLFVBQVUsQ0FBQyxXQUFXLEVBQUU7U0FDbEQsQ0FBQztRQUVGLGtDQUFrQztRQUNsQyxNQUFNLFVBQVUsR0FBRyxNQUFNLElBQUksQ0FBQyxlQUFlLENBQUMsSUFBSSxDQUFDLFdBQVcsQ0FBQyxNQUFNLENBQUM7WUFDcEUsS0FBSyxFQUFFLGVBQWU7WUFDdEIsUUFBUSxFQUFFLFFBQVE7WUFDbEIsTUFBTSxFQUFFLEtBQUs7U0FDZCxDQUFDLENBQUM7UUFFSCxrQ0FBa0M7UUFDbEMsT0FBTztZQUNMLElBQUksRUFBRSxXQUFXO1lBQ2pCLE9BQU8sRUFBRSxVQUFVLENBQUMsT0FBTyxDQUFDLENBQUMsQ0FBQyxFQUFFLE9BQU8sRUFBRSxPQUFPLElBQUksRUFBRTtTQUN2RCxDQUFDO0lBQ0osQ0FBQztJQUVNLEtBQUssQ0FBQyxLQUFLLENBQUMsVUFBK0I7UUFDaEQsTUFBTSxJQUFJLEtBQUssQ0FBQywyQ0FBMkMsQ0FBQyxDQUFDO0lBQy9ELENBQUM7SUFFTSxLQUFLLENBQUMsTUFBTSxDQUFDLFVBQTZDO1FBQy9ELE1BQU0sSUFBSSxLQUFLLENBQUMsd0NBQXdDLENBQUMsQ0FBQztJQUM1RCxDQUFDO0lBRU0sS0FBSyxDQUFDLFFBQVEsQ0FBQyxVQUtyQjtRQUNDLHdDQUF3QztRQUN4QyxJQUFJLDBCQUEwQixHQUFpQixFQUFFLENBQUM7UUFFbEQsS0FBSyxNQUFNLFdBQVcsSUFBSSxVQUFVLENBQUMsWUFBWSxFQUFFLENBQUM7WUFDbEQsTUFBTSxrQkFBa0IsR0FBRyxNQUFNLElBQUksQ0FBQyxnQkFBZ0IsQ0FBQyxvQkFBb0IsQ0FBQyxXQUFXLENBQUMsQ0FBQztZQUN6RiwwQkFBMEIsR0FBRywwQkFBMEIsQ0FBQyxNQUFNLENBQUMsa0JBQWtCLENBQUMsQ0FBQztRQUNyRixDQUFDO1FBRUQsd0RBQXdEO1FBQ3hELE1BQU0sZ0JBQWdCLEdBQUcsMEJBQTBCLENBQUMsR0FBRyxDQUFDLEtBQUssQ0FBQyxFQUFFLENBQzlELE1BQU0sQ0FBQyxJQUFJLENBQUMsS0FBSyxDQUFDLENBQUMsUUFBUSxDQUFDLFFBQVEsQ0FBQyxDQUN0QyxDQUFDO1FBRUYsZ0RBQWdEO1FBQ2hELE1BQU0sbUJBQW1CLEdBQUc7UUFDeEIsVUFBVSxDQUFDLFdBQVc7OztRQUd0QixnQkFBZ0IsQ0FBQyxHQUFHLENBQUMsQ0FBQyxHQUFHLEVBQUUsQ0FBQyxFQUFFLEVBQUUsQ0FBQyxTQUFTLENBQUMsR0FBRyxDQUFDLGdCQUFnQixDQUFDLENBQUMsSUFBSSxDQUFDLElBQUksQ0FBQztLQUM5RSxDQUFDO1FBRUYsK0NBQStDO1FBQy9DLE1BQU0sUUFBUSxHQUFpQztZQUM3QyxFQUFFLElBQUksRUFBRSxRQUFRLEVBQUUsT0FBTyxFQUFFLFVBQVUsQ0FBQyxhQUFhLEVBQUU7WUFDckQsR0FBRyxVQUFVLENBQUMsY0FBYyxDQUFDLEdBQUcsQ0FBQyxHQUFHLENBQUMsRUFBRSxDQUFDLENBQUM7Z0JBQ3ZDLElBQUksRUFBRSxHQUFHLENBQUMsSUFBdUM7Z0JBQ2pELE9BQU8sRUFBRSxHQUFHLENBQUMsT0FBTzthQUNyQixDQUFDLENBQUM7WUFDSCxFQUFFLElBQUksRUFBRSxNQUFNLEVBQUUsT0FBTyxFQUFFLG1CQUFtQixFQUFFO1NBQy9DLENBQUM7UUFFRixNQUFNLFVBQVUsR0FBRyxNQUFNLElBQUksQ0FBQyxlQUFlLENBQUMsSUFBSSxDQUFDLFdBQVcsQ0FBQyxNQUFNLENBQUM7WUFDcEUsS0FBSyxFQUFFLGVBQWU7WUFDdEIsUUFBUSxFQUFFLFFBQVE7WUFDbEIsTUFBTSxFQUFFLEtBQUs7U0FDZCxDQUFDLENBQUM7UUFFSCxPQUFPO1lBQ0wsT0FBTyxFQUFFLFVBQVUsQ0FBQyxPQUFPLENBQUMsQ0FBQyxDQUFDLEVBQUUsT0FBTyxFQUFFLE9BQU8sSUFBSSxFQUFFO1NBQ3ZELENBQUM7SUFDSixDQUFDO0NBQ0YifQ==
package/license ADDED
@@ -0,0 +1,19 @@
1
+ Copyright (c) 2024 Task Venture Capital GmbH (hello@task.vc)
2
+
3
+ Permission is hereby granted, free of charge, to any person obtaining a copy
4
+ of this software and associated documentation files (the "Software"), to deal
5
+ in the Software without restriction, including without limitation the rights
6
+ to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
7
+ copies of the Software, and to permit persons to whom the Software is
8
+ furnished to do so, subject to the following conditions:
9
+
10
+ The above copyright notice and this permission notice shall be included in all
11
+ copies or substantial portions of the Software.
12
+
13
+ THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14
+ IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15
+ FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
16
+ AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
17
+ LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
18
+ OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
19
+ SOFTWARE.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@push.rocks/smartai",
3
- "version": "0.2.0",
3
+ "version": "0.3.0",
4
4
  "private": false,
5
5
  "description": "A TypeScript library for integrating and interacting with multiple AI models, offering capabilities for chat and potentially audio responses.",
6
6
  "main": "dist_ts/index.js",
package/readme.md CHANGED
@@ -24,6 +24,14 @@ This command installs the package and adds it to your project's dependencies.
24
24
  openaiToken: 'your-openai-token'
25
25
  ```
26
26
 
27
+ ### X.AI
28
+ - Models: Grok-2-latest
29
+ - Features: Chat, Streaming, Document Processing
30
+ - Configuration:
31
+ ```typescript
32
+ xaiToken: 'your-xai-token'
33
+ ```
34
+
27
35
  ### Anthropic
28
36
  - Models: Claude-3-opus-20240229
29
37
  - Features: Chat, Streaming, Vision, Document Processing
@@ -75,6 +83,7 @@ import { SmartAi } from '@push.rocks/smartai';
75
83
 
76
84
  const smartAi = new SmartAi({
77
85
  openaiToken: 'your-openai-token',
86
+ xaiToken: 'your-xai-token',
78
87
  anthropicToken: 'your-anthropic-token',
79
88
  perplexityToken: 'your-perplexity-token',
80
89
  groqToken: 'your-groq-token',
@@ -3,6 +3,6 @@
3
3
  */
4
4
  export const commitinfo = {
5
5
  name: '@push.rocks/smartai',
6
- version: '0.2.0',
6
+ version: '0.3.0',
7
7
  description: 'A TypeScript library for integrating and interacting with multiple AI models, offering capabilities for chat and potentially audio responses.'
8
8
  }
@@ -0,0 +1,183 @@
1
+ import * as plugins from './plugins.js';
2
+ import * as paths from './paths.js';
3
+ import { MultiModalModel } from './abstract.classes.multimodal.js';
4
+ import type { ChatOptions, ChatResponse, ChatMessage } from './abstract.classes.multimodal.js';
5
+ import type { ChatCompletionMessageParam } from 'openai/resources/chat/completions';
6
+
7
+ export interface IXAIProviderOptions {
8
+ xaiToken: string;
9
+ }
10
+
11
+ export class XAIProvider extends MultiModalModel {
12
+ private options: IXAIProviderOptions;
13
+ public openAiApiClient: plugins.openai.default;
14
+ public smartpdfInstance: plugins.smartpdf.SmartPdf;
15
+
16
+ constructor(optionsArg: IXAIProviderOptions) {
17
+ super();
18
+ this.options = optionsArg;
19
+ }
20
+
21
+ public async start() {
22
+ this.openAiApiClient = new plugins.openai.default({
23
+ apiKey: this.options.xaiToken,
24
+ baseURL: 'https://api.x.ai/v1',
25
+ });
26
+ this.smartpdfInstance = new plugins.smartpdf.SmartPdf();
27
+ }
28
+
29
+ public async stop() {}
30
+
31
+ public async chatStream(input: ReadableStream<Uint8Array>): Promise<ReadableStream<string>> {
32
+ // Create a TextDecoder to handle incoming chunks
33
+ const decoder = new TextDecoder();
34
+ let buffer = '';
35
+ let currentMessage: { role: string; content: string; } | null = null;
36
+
37
+ // Create a TransformStream to process the input
38
+ const transform = new TransformStream<Uint8Array, string>({
39
+ async transform(chunk, controller) {
40
+ buffer += decoder.decode(chunk, { stream: true });
41
+
42
+ // Try to parse complete JSON messages from the buffer
43
+ while (true) {
44
+ const newlineIndex = buffer.indexOf('\n');
45
+ if (newlineIndex === -1) break;
46
+
47
+ const line = buffer.slice(0, newlineIndex);
48
+ buffer = buffer.slice(newlineIndex + 1);
49
+
50
+ if (line.trim()) {
51
+ try {
52
+ const message = JSON.parse(line);
53
+ currentMessage = {
54
+ role: message.role || 'user',
55
+ content: message.content || '',
56
+ };
57
+ } catch (e) {
58
+ console.error('Failed to parse message:', e);
59
+ }
60
+ }
61
+ }
62
+
63
+ // If we have a complete message, send it to X.AI
64
+ if (currentMessage) {
65
+ const stream = await this.openAiApiClient.chat.completions.create({
66
+ model: 'grok-2-latest',
67
+ messages: [{ role: currentMessage.role, content: currentMessage.content }],
68
+ stream: true,
69
+ });
70
+
71
+ // Process each chunk from X.AI
72
+ for await (const chunk of stream) {
73
+ const content = chunk.choices[0]?.delta?.content;
74
+ if (content) {
75
+ controller.enqueue(content);
76
+ }
77
+ }
78
+
79
+ currentMessage = null;
80
+ }
81
+ },
82
+
83
+ flush(controller) {
84
+ if (buffer) {
85
+ try {
86
+ const message = JSON.parse(buffer);
87
+ controller.enqueue(message.content || '');
88
+ } catch (e) {
89
+ console.error('Failed to parse remaining buffer:', e);
90
+ }
91
+ }
92
+ }
93
+ });
94
+
95
+ // Connect the input to our transform stream
96
+ return input.pipeThrough(transform);
97
+ }
98
+
99
+ public async chat(optionsArg: {
100
+ systemMessage: string;
101
+ userMessage: string;
102
+ messageHistory: { role: string; content: string; }[];
103
+ }): Promise<{ role: 'assistant'; message: string; }> {
104
+ // Prepare messages array with system message, history, and user message
105
+ const messages: ChatCompletionMessageParam[] = [
106
+ { role: 'system', content: optionsArg.systemMessage },
107
+ ...optionsArg.messageHistory.map(msg => ({
108
+ role: msg.role as 'system' | 'user' | 'assistant',
109
+ content: msg.content
110
+ })),
111
+ { role: 'user', content: optionsArg.userMessage }
112
+ ];
113
+
114
+ // Call X.AI's chat completion API
115
+ const completion = await this.openAiApiClient.chat.completions.create({
116
+ model: 'grok-2-latest',
117
+ messages: messages,
118
+ stream: false,
119
+ });
120
+
121
+ // Return the assistant's response
122
+ return {
123
+ role: 'assistant',
124
+ message: completion.choices[0]?.message?.content || ''
125
+ };
126
+ }
127
+
128
+ public async audio(optionsArg: { message: string }): Promise<NodeJS.ReadableStream> {
129
+ throw new Error('Audio generation is not supported by X.AI');
130
+ }
131
+
132
+ public async vision(optionsArg: { image: Buffer; prompt: string }): Promise<string> {
133
+ throw new Error('Vision tasks are not supported by X.AI');
134
+ }
135
+
136
+ public async document(optionsArg: {
137
+ systemMessage: string;
138
+ userMessage: string;
139
+ pdfDocuments: Uint8Array[];
140
+ messageHistory: { role: string; content: string; }[];
141
+ }): Promise<{ message: any }> {
142
+ // First convert PDF documents to images
143
+ let pdfDocumentImageBytesArray: Uint8Array[] = [];
144
+
145
+ for (const pdfDocument of optionsArg.pdfDocuments) {
146
+ const documentImageArray = await this.smartpdfInstance.convertPDFToPngBytes(pdfDocument);
147
+ pdfDocumentImageBytesArray = pdfDocumentImageBytesArray.concat(documentImageArray);
148
+ }
149
+
150
+ // Convert images to base64 for inclusion in the message
151
+ const imageBase64Array = pdfDocumentImageBytesArray.map(bytes =>
152
+ Buffer.from(bytes).toString('base64')
153
+ );
154
+
155
+ // Combine document images into the user message
156
+ const enhancedUserMessage = `
157
+ ${optionsArg.userMessage}
158
+
159
+ Document contents (as images):
160
+ ${imageBase64Array.map((img, i) => `Image ${i + 1}: <image data>`).join('\n')}
161
+ `;
162
+
163
+ // Use chat completion to analyze the documents
164
+ const messages: ChatCompletionMessageParam[] = [
165
+ { role: 'system', content: optionsArg.systemMessage },
166
+ ...optionsArg.messageHistory.map(msg => ({
167
+ role: msg.role as 'system' | 'user' | 'assistant',
168
+ content: msg.content
169
+ })),
170
+ { role: 'user', content: enhancedUserMessage }
171
+ ];
172
+
173
+ const completion = await this.openAiApiClient.chat.completions.create({
174
+ model: 'grok-2-latest',
175
+ messages: messages,
176
+ stream: false,
177
+ });
178
+
179
+ return {
180
+ message: completion.choices[0]?.message?.content || ''
181
+ };
182
+ }
183
+ }