@push.rocks/smartai 0.1.0 → 0.2.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist_ts/00_commitinfo_data.js +1 -1
- package/dist_ts/provider.anthropic.js +84 -3
- package/package.json +1 -1
- package/readme.md +17 -3
- package/ts/00_commitinfo_data.ts +1 -1
- package/ts/provider.anthropic.ts +96 -2
|
@@ -3,7 +3,7 @@
|
|
|
3
3
|
*/
|
|
4
4
|
export const commitinfo = {
|
|
5
5
|
name: '@push.rocks/smartai',
|
|
6
|
-
version: '0.
|
|
6
|
+
version: '0.2.0',
|
|
7
7
|
description: 'A TypeScript library for integrating and interacting with multiple AI models, offering capabilities for chat and potentially audio responses.'
|
|
8
8
|
};
|
|
9
9
|
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiMDBfY29tbWl0aW5mb19kYXRhLmpzIiwic291cmNlUm9vdCI6IiIsInNvdXJjZXMiOlsiLi4vdHMvMDBfY29tbWl0aW5mb19kYXRhLnRzIl0sIm5hbWVzIjpbXSwibWFwcGluZ3MiOiJBQUFBOztHQUVHO0FBQ0gsTUFBTSxDQUFDLE1BQU0sVUFBVSxHQUFHO0lBQ3hCLElBQUksRUFBRSxxQkFBcUI7SUFDM0IsT0FBTyxFQUFFLE9BQU87SUFDaEIsV0FBVyxFQUFFLCtJQUErSTtDQUM3SixDQUFBIn0=
|
|
@@ -108,10 +108,91 @@ export class AnthropicProvider extends MultiModalModel {
|
|
|
108
108
|
throw new Error('Audio generation is not yet supported by Anthropic.');
|
|
109
109
|
}
|
|
110
110
|
async vision(optionsArg) {
|
|
111
|
-
|
|
111
|
+
const base64Image = optionsArg.image.toString('base64');
|
|
112
|
+
const content = [
|
|
113
|
+
{
|
|
114
|
+
type: 'text',
|
|
115
|
+
text: optionsArg.prompt
|
|
116
|
+
},
|
|
117
|
+
{
|
|
118
|
+
type: 'image',
|
|
119
|
+
source: {
|
|
120
|
+
type: 'base64',
|
|
121
|
+
media_type: 'image/jpeg',
|
|
122
|
+
data: base64Image
|
|
123
|
+
}
|
|
124
|
+
}
|
|
125
|
+
];
|
|
126
|
+
const result = await this.anthropicApiClient.messages.create({
|
|
127
|
+
model: 'claude-3-opus-20240229',
|
|
128
|
+
messages: [{
|
|
129
|
+
role: 'user',
|
|
130
|
+
content
|
|
131
|
+
}],
|
|
132
|
+
max_tokens: 1024
|
|
133
|
+
});
|
|
134
|
+
// Extract text content from the response
|
|
135
|
+
let message = '';
|
|
136
|
+
for (const block of result.content) {
|
|
137
|
+
if ('text' in block) {
|
|
138
|
+
message += block.text;
|
|
139
|
+
}
|
|
140
|
+
}
|
|
141
|
+
return message;
|
|
112
142
|
}
|
|
113
143
|
async document(optionsArg) {
|
|
114
|
-
|
|
144
|
+
// Convert PDF documents to images using SmartPDF
|
|
145
|
+
const smartpdfInstance = new plugins.smartpdf.SmartPdf();
|
|
146
|
+
let documentImageBytesArray = [];
|
|
147
|
+
for (const pdfDocument of optionsArg.pdfDocuments) {
|
|
148
|
+
const documentImageArray = await smartpdfInstance.convertPDFToPngBytes(pdfDocument);
|
|
149
|
+
documentImageBytesArray = documentImageBytesArray.concat(documentImageArray);
|
|
150
|
+
}
|
|
151
|
+
// Convert message history to Anthropic format
|
|
152
|
+
const messages = optionsArg.messageHistory.map(msg => ({
|
|
153
|
+
role: msg.role === 'assistant' ? 'assistant' : 'user',
|
|
154
|
+
content: msg.content
|
|
155
|
+
}));
|
|
156
|
+
// Create content array with text and images
|
|
157
|
+
const content = [
|
|
158
|
+
{
|
|
159
|
+
type: 'text',
|
|
160
|
+
text: optionsArg.userMessage
|
|
161
|
+
}
|
|
162
|
+
];
|
|
163
|
+
// Add each document page as an image
|
|
164
|
+
for (const imageBytes of documentImageBytesArray) {
|
|
165
|
+
content.push({
|
|
166
|
+
type: 'image',
|
|
167
|
+
source: {
|
|
168
|
+
type: 'base64',
|
|
169
|
+
media_type: 'image/jpeg',
|
|
170
|
+
data: Buffer.from(imageBytes).toString('base64')
|
|
171
|
+
}
|
|
172
|
+
});
|
|
173
|
+
}
|
|
174
|
+
const result = await this.anthropicApiClient.messages.create({
|
|
175
|
+
model: 'claude-3-opus-20240229',
|
|
176
|
+
system: optionsArg.systemMessage,
|
|
177
|
+
messages: [
|
|
178
|
+
...messages,
|
|
179
|
+
{ role: 'user', content }
|
|
180
|
+
],
|
|
181
|
+
max_tokens: 4096
|
|
182
|
+
});
|
|
183
|
+
// Extract text content from the response
|
|
184
|
+
let message = '';
|
|
185
|
+
for (const block of result.content) {
|
|
186
|
+
if ('text' in block) {
|
|
187
|
+
message += block.text;
|
|
188
|
+
}
|
|
189
|
+
}
|
|
190
|
+
return {
|
|
191
|
+
message: {
|
|
192
|
+
role: 'assistant',
|
|
193
|
+
content: message
|
|
194
|
+
}
|
|
195
|
+
};
|
|
115
196
|
}
|
|
116
197
|
}
|
|
117
|
-
//# sourceMappingURL=data:application/json;base64,
|
|
198
|
+
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoicHJvdmlkZXIuYW50aHJvcGljLmpzIiwic291cmNlUm9vdCI6IiIsInNvdXJjZXMiOlsiLi4vdHMvcHJvdmlkZXIuYW50aHJvcGljLnRzIl0sIm5hbWVzIjpbXSwibWFwcGluZ3MiOiJBQUFBLE9BQU8sS0FBSyxPQUFPLE1BQU0sY0FBYyxDQUFDO0FBQ3hDLE9BQU8sS0FBSyxLQUFLLE1BQU0sWUFBWSxDQUFDO0FBQ3BDLE9BQU8sRUFBRSxlQUFlLEVBQUUsTUFBTSxrQ0FBa0MsQ0FBQztBQVVuRSxNQUFNLE9BQU8saUJBQWtCLFNBQVEsZUFBZTtJQUlwRCxZQUFZLFVBQXFDO1FBQy9DLEtBQUssRUFBRSxDQUFDO1FBQ1IsSUFBSSxDQUFDLE9BQU8sR0FBRyxVQUFVLENBQUEsQ0FBQyw2QkFBNkI7SUFDekQsQ0FBQztJQUVELEtBQUssQ0FBQyxLQUFLO1FBQ1QsSUFBSSxDQUFDLGtCQUFrQixHQUFHLElBQUksT0FBTyxDQUFDLFNBQVMsQ0FBQyxPQUFPLENBQUM7WUFDdEQsTUFBTSxFQUFFLElBQUksQ0FBQyxPQUFPLENBQUMsY0FBYztTQUNwQyxDQUFDLENBQUM7SUFDTCxDQUFDO0lBRUQsS0FBSyxDQUFDLElBQUksS0FBSSxDQUFDO0lBRVIsS0FBSyxDQUFDLFVBQVUsQ0FBQyxLQUFpQztRQUN2RCxpREFBaUQ7UUFDakQsTUFBTSxPQUFPLEdBQUcsSUFBSSxXQUFXLEVBQUUsQ0FBQztRQUNsQyxJQUFJLE1BQU0sR0FBRyxFQUFFLENBQUM7UUFDaEIsSUFBSSxjQUFjLEdBQThDLElBQUksQ0FBQztRQUVyRSxnREFBZ0Q7UUFDaEQsTUFBTSxTQUFTLEdBQUcsSUFBSSxlQUFlLENBQXFCO1lBQ3hELEtBQUssQ0FBQyxTQUFTLENBQUMsS0FBSyxFQUFFLFVBQVU7Z0JBQy9CLE1BQU0sSUFBSSxPQUFPLENBQUMsTUFBTSxDQUFDLEtBQUssRUFBRSxFQUFFLE1BQU0sRUFBRSxJQUFJLEVBQUUsQ0FBQyxDQUFDO2dCQUVsRCxzREFBc0Q7Z0JBQ3RELE9BQU8sSUFBSSxFQUFFLENBQUM7b0JBQ1osTUFBTSxZQUFZLEdBQUcsTUFBTSxDQUFDLE9BQU8sQ0FBQyxJQUFJLENBQUMsQ0FBQztvQkFDMUMsSUFBSSxZQUFZLEtBQUssQ0FBQyxDQUFDO3dCQUFFLE1BQU07b0JBRS9CLE1BQU0sSUFBSSxHQUFHLE1BQU0sQ0FBQyxLQUFLLENBQUMsQ0FBQyxFQUFFLFlBQVksQ0FBQyxDQUFDO29CQUMzQyxNQUFNLEdBQUcsTUFBTSxDQUFDLEtBQUssQ0FBQyxZQUFZLEdBQUcsQ0FBQyxDQUFDLENBQUM7b0JBRXhDLElBQUksSUFBSSxDQUFDLElBQUksRUFBRSxFQUFFLENBQUM7d0JBQ2hCLElBQUksQ0FBQzs0QkFDSCxNQUFNLE9BQU8sR0FBRyxJQUFJLENBQUMsS0FBSyxDQUFDLElBQUksQ0FBQyxDQUFDOzRCQUNqQyxjQUFjLEdBQUc7Z0NBQ2YsSUFBSSxFQUFFLE9BQU8sQ0FBQyxJQUFJLElBQUksTUFBTTtnQ0FDNUIsT0FBTyxFQUFFLE9BQU8sQ0FBQyxPQUFPLElBQUksRUFBRTs2QkFDL0IsQ0FBQzt3QkFDSixDQUFDO3dCQUFDLE9BQU8sQ0FBQyxFQUFFLENBQUM7NEJBQ1gsT0FBTyxDQUFDLEtBQUssQ0FBQywwQkFBMEIsRUFBRSxDQUFDLENBQUMsQ0FBQzt3QkFDL0MsQ0FBQztvQkFDSCxDQUFDO2dCQUNILENBQUM7Z0JBRUQsc0RBQXNEO2dCQUN0RCxJQUFJLGNBQWMsRUFBRSxDQUFDO29CQUNuQixNQUFNLE1BQU0sR0FBRyxNQUFNLElBQUksQ0FBQyxrQkFBa0IsQ0FBQyxRQUFRLENBQUMsTUFBTSxDQUFDO3dCQUMzRCxLQUFLLEVBQUUsd0JBQXdCO3dCQUMvQixRQUFRLEVBQUUsQ0FBQyxFQUFFLElBQUksRUFBRSxjQUFjLENBQUMsSUFBSSxFQUFFLE9BQU8sRUFBRSxjQUFjLENBQUMsT0FBTyxFQUFFLENBQUM7d0JBQzFFLE1BQU0sRUFBRSxFQUFFO3dCQUNWLE1BQU0sRUFBRSxJQUFJO3dCQUNaLFVBQVUsRUFBRSxJQUFJO3FCQUNqQixDQUFDLENBQUM7b0JBRUgsb0NBQW9DO29CQUNwQyxJQUFJLEtBQUssRUFBRSxNQUFNLEtBQUssSUFBSSxNQUFNLEVBQUUsQ0FBQzt3QkFDakMsTUFBTSxPQUFPLEdBQUcsS0FBSyxDQUFDLEtBQUssRUFBRSxJQUFJLENBQUM7d0JBQ2xDLElBQUksT0FBTyxFQUFFLENBQUM7NEJBQ1osVUFBVSxDQUFDLE9BQU8sQ0FBQyxPQUFPLENBQUMsQ0FBQzt3QkFDOUIsQ0FBQztvQkFDSCxDQUFDO29CQUVELGNBQWMsR0FBRyxJQUFJLENBQUM7Z0JBQ3hCLENBQUM7WUFDSCxDQUFDO1lBRUQsS0FBSyxDQUFDLFVBQVU7Z0JBQ2QsSUFBSSxNQUFNLEVBQUUsQ0FBQztvQkFDWCxJQUFJLENBQUM7d0JBQ0gsTUFBTSxPQUFPLEdBQUcsSUFBSSxDQUFDLEtBQUssQ0FBQyxNQUFNLENBQUMsQ0FBQzt3QkFDbkMsVUFBVSxDQUFDLE9BQU8sQ0FBQyxPQUFPLENBQUMsT0FBTyxJQUFJLEVBQUUsQ0FBQyxDQUFDO29CQUM1QyxDQUFDO29CQUFDLE9BQU8sQ0FBQyxFQUFFLENBQUM7d0JBQ1gsT0FBTyxDQUFDLEtBQUssQ0FBQyxtQ0FBbUMsRUFBRSxDQUFDLENBQUMsQ0FBQztvQkFDeEQsQ0FBQztnQkFDSCxDQUFDO1lBQ0gsQ0FBQztTQUNGLENBQUMsQ0FBQztRQUVILDRDQUE0QztRQUM1QyxPQUFPLEtBQUssQ0FBQyxXQUFXLENBQUMsU0FBUyxDQUFDLENBQUM7SUFDdEMsQ0FBQztJQUVELGdEQUFnRDtJQUN6QyxLQUFLLENBQUMsSUFBSSxDQUFDLFVBQXVCO1FBQ3ZDLDhDQUE4QztRQUM5QyxNQUFNLFFBQVEsR0FBRyxVQUFVLENBQUMsY0FBYyxDQUFDLEdBQUcsQ0FBQyxHQUFHLENBQUMsRUFBRSxDQUFDLENBQUM7WUFDckQsSUFBSSxFQUFFLEdBQUcsQ0FBQyxJQUFJLEtBQUssV0FBVyxDQUFDLENBQUMsQ0FBQyxXQUFvQixDQUFDLENBQUMsQ0FBQyxNQUFlO1lBQ3ZFLE9BQU8sRUFBRSxHQUFHLENBQUMsT0FBTztTQUNyQixDQUFDLENBQUMsQ0FBQztRQUVKLE1BQU0sTUFBTSxHQUFHLE1BQU0sSUFBSSxDQUFDLGtCQUFrQixDQUFDLFFBQVEsQ0FBQyxNQUFNLENBQUM7WUFDM0QsS0FBSyxFQUFFLHdCQUF3QjtZQUMvQixNQUFNLEVBQUUsVUFBVSxDQUFDLGFBQWE7WUFDaEMsUUFBUSxFQUFFO2dCQUNSLEdBQUcsUUFBUTtnQkFDWCxFQUFFLElBQUksRUFBRSxNQUFlLEVBQUUsT0FBTyxFQUFFLFVBQVUsQ0FBQyxXQUFXLEVBQUU7YUFDM0Q7WUFDRCxVQUFVLEVBQUUsSUFBSTtTQUNqQixDQUFDLENBQUM7UUFFSCx5Q0FBeUM7UUFDekMsSUFBSSxPQUFPLEdBQUcsRUFBRSxDQUFDO1FBQ2pCLEtBQUssTUFBTSxLQUFLLElBQUksTUFBTSxDQUFDLE9BQU8sRUFBRSxDQUFDO1lBQ25DLElBQUksTUFBTSxJQUFJLEtBQUssRUFBRSxDQUFDO2dCQUNwQixPQUFPLElBQUksS0FBSyxDQUFDLElBQUksQ0FBQztZQUN4QixDQUFDO1FBQ0gsQ0FBQztRQUVELE9BQU87WUFDTCxJQUFJLEVBQUUsV0FBb0I7WUFDMUIsT0FBTztTQUNSLENBQUM7SUFDSixDQUFDO0lBRU0sS0FBSyxDQUFDLEtBQUssQ0FBQyxVQUErQjtRQUNoRCw4RUFBOEU7UUFDOUUsTUFBTSxJQUFJLEtBQUssQ0FBQyxxREFBcUQsQ0FBQyxDQUFDO0lBQ3pFLENBQUM7SUFFTSxLQUFLLENBQUMsTUFBTSxDQUFDLFVBQTZDO1FBQy9ELE1BQU0sV0FBVyxHQUFHLFVBQVUsQ0FBQyxLQUFLLENBQUMsUUFBUSxDQUFDLFFBQVEsQ0FBQyxDQUFDO1FBRXhELE1BQU0sT0FBTyxHQUFtQjtZQUM5QjtnQkFDRSxJQUFJLEVBQUUsTUFBTTtnQkFDWixJQUFJLEVBQUUsVUFBVSxDQUFDLE1BQU07YUFDeEI7WUFDRDtnQkFDRSxJQUFJLEVBQUUsT0FBTztnQkFDYixNQUFNLEVBQUU7b0JBQ04sSUFBSSxFQUFFLFFBQVE7b0JBQ2QsVUFBVSxFQUFFLFlBQVk7b0JBQ3hCLElBQUksRUFBRSxXQUFXO2lCQUNsQjthQUNGO1NBQ0YsQ0FBQztRQUVGLE1BQU0sTUFBTSxHQUFHLE1BQU0sSUFBSSxDQUFDLGtCQUFrQixDQUFDLFFBQVEsQ0FBQyxNQUFNLENBQUM7WUFDM0QsS0FBSyxFQUFFLHdCQUF3QjtZQUMvQixRQUFRLEVBQUUsQ0FBQztvQkFDVCxJQUFJLEVBQUUsTUFBTTtvQkFDWixPQUFPO2lCQUNSLENBQUM7WUFDRixVQUFVLEVBQUUsSUFBSTtTQUNqQixDQUFDLENBQUM7UUFFSCx5Q0FBeUM7UUFDekMsSUFBSSxPQUFPLEdBQUcsRUFBRSxDQUFDO1FBQ2pCLEtBQUssTUFBTSxLQUFLLElBQUksTUFBTSxDQUFDLE9BQU8sRUFBRSxDQUFDO1lBQ25DLElBQUksTUFBTSxJQUFJLEtBQUssRUFBRSxDQUFDO2dCQUNwQixPQUFPLElBQUksS0FBSyxDQUFDLElBQUksQ0FBQztZQUN4QixDQUFDO1FBQ0gsQ0FBQztRQUNELE9BQU8sT0FBTyxDQUFDO0lBQ2pCLENBQUM7SUFFTSxLQUFLLENBQUMsUUFBUSxDQUFDLFVBS3JCO1FBQ0MsaURBQWlEO1FBQ2pELE1BQU0sZ0JBQWdCLEdBQUcsSUFBSSxPQUFPLENBQUMsUUFBUSxDQUFDLFFBQVEsRUFBRSxDQUFDO1FBQ3pELElBQUksdUJBQXVCLEdBQWlCLEVBQUUsQ0FBQztRQUUvQyxLQUFLLE1BQU0sV0FBVyxJQUFJLFVBQVUsQ0FBQyxZQUFZLEVBQUUsQ0FBQztZQUNsRCxNQUFNLGtCQUFrQixHQUFHLE1BQU0sZ0JBQWdCLENBQUMsb0JBQW9CLENBQUMsV0FBVyxDQUFDLENBQUM7WUFDcEYsdUJBQXVCLEdBQUcsdUJBQXVCLENBQUMsTUFBTSxDQUFDLGtCQUFrQixDQUFDLENBQUM7UUFDL0UsQ0FBQztRQUVELDhDQUE4QztRQUM5QyxNQUFNLFFBQVEsR0FBRyxVQUFVLENBQUMsY0FBYyxDQUFDLEdBQUcsQ0FBQyxHQUFHLENBQUMsRUFBRSxDQUFDLENBQUM7WUFDckQsSUFBSSxFQUFFLEdBQUcsQ0FBQyxJQUFJLEtBQUssV0FBVyxDQUFDLENBQUMsQ0FBQyxXQUFvQixDQUFDLENBQUMsQ0FBQyxNQUFlO1lBQ3ZFLE9BQU8sRUFBRSxHQUFHLENBQUMsT0FBTztTQUNyQixDQUFDLENBQUMsQ0FBQztRQUVKLDRDQUE0QztRQUM1QyxNQUFNLE9BQU8sR0FBbUI7WUFDOUI7Z0JBQ0UsSUFBSSxFQUFFLE1BQU07Z0JBQ1osSUFBSSxFQUFFLFVBQVUsQ0FBQyxXQUFXO2FBQzdCO1NBQ0YsQ0FBQztRQUVGLHFDQUFxQztRQUNyQyxLQUFLLE1BQU0sVUFBVSxJQUFJLHVCQUF1QixFQUFFLENBQUM7WUFDakQsT0FBTyxDQUFDLElBQUksQ0FBQztnQkFDWCxJQUFJLEVBQUUsT0FBTztnQkFDYixNQUFNLEVBQUU7b0JBQ04sSUFBSSxFQUFFLFFBQVE7b0JBQ2QsVUFBVSxFQUFFLFlBQVk7b0JBQ3hCLElBQUksRUFBRSxNQUFNLENBQUMsSUFBSSxDQUFDLFVBQVUsQ0FBQyxDQUFDLFFBQVEsQ0FBQyxRQUFRLENBQUM7aUJBQ2pEO2FBQ0YsQ0FBQyxDQUFDO1FBQ0wsQ0FBQztRQUVELE1BQU0sTUFBTSxHQUFHLE1BQU0sSUFBSSxDQUFDLGtCQUFrQixDQUFDLFFBQVEsQ0FBQyxNQUFNLENBQUM7WUFDM0QsS0FBSyxFQUFFLHdCQUF3QjtZQUMvQixNQUFNLEVBQUUsVUFBVSxDQUFDLGFBQWE7WUFDaEMsUUFBUSxFQUFFO2dCQUNSLEdBQUcsUUFBUTtnQkFDWCxFQUFFLElBQUksRUFBRSxNQUFNLEVBQUUsT0FBTyxFQUFFO2FBQzFCO1lBQ0QsVUFBVSxFQUFFLElBQUk7U0FDakIsQ0FBQyxDQUFDO1FBRUgseUNBQXlDO1FBQ3pDLElBQUksT0FBTyxHQUFHLEVBQUUsQ0FBQztRQUNqQixLQUFLLE1BQU0sS0FBSyxJQUFJLE1BQU0sQ0FBQyxPQUFPLEVBQUUsQ0FBQztZQUNuQyxJQUFJLE1BQU0sSUFBSSxLQUFLLEVBQUUsQ0FBQztnQkFDcEIsT0FBTyxJQUFJLEtBQUssQ0FBQyxJQUFJLENBQUM7WUFDeEIsQ0FBQztRQUNILENBQUM7UUFFRCxPQUFPO1lBQ0wsT0FBTyxFQUFFO2dCQUNQLElBQUksRUFBRSxXQUFXO2dCQUNqQixPQUFPLEVBQUUsT0FBTzthQUNqQjtTQUNGLENBQUM7SUFDSixDQUFDO0NBQ0YifQ==
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@push.rocks/smartai",
|
|
3
|
-
"version": "0.
|
|
3
|
+
"version": "0.2.0",
|
|
4
4
|
"private": false,
|
|
5
5
|
"description": "A TypeScript library for integrating and interacting with multiple AI models, offering capabilities for chat and potentially audio responses.",
|
|
6
6
|
"main": "dist_ts/index.js",
|
package/readme.md
CHANGED
|
@@ -26,7 +26,7 @@ This command installs the package and adds it to your project's dependencies.
|
|
|
26
26
|
|
|
27
27
|
### Anthropic
|
|
28
28
|
- Models: Claude-3-opus-20240229
|
|
29
|
-
- Features: Chat, Streaming
|
|
29
|
+
- Features: Chat, Streaming, Vision, Document Processing
|
|
30
30
|
- Configuration:
|
|
31
31
|
```typescript
|
|
32
32
|
anthropicToken: 'your-anthropic-token'
|
|
@@ -148,7 +148,7 @@ const audioStream = await smartAi.openaiProvider.audio({
|
|
|
148
148
|
|
|
149
149
|
### Document Processing
|
|
150
150
|
|
|
151
|
-
For providers that support document processing (OpenAI and
|
|
151
|
+
For providers that support document processing (OpenAI, Ollama, and Anthropic):
|
|
152
152
|
|
|
153
153
|
```typescript
|
|
154
154
|
// Using OpenAI
|
|
@@ -166,6 +166,14 @@ const analysis = await smartAi.ollamaProvider.document({
|
|
|
166
166
|
messageHistory: [],
|
|
167
167
|
pdfDocuments: [pdfBuffer] // Uint8Array of PDF content
|
|
168
168
|
});
|
|
169
|
+
|
|
170
|
+
// Using Anthropic with Claude 3
|
|
171
|
+
const anthropicAnalysis = await smartAi.anthropicProvider.document({
|
|
172
|
+
systemMessage: 'You are a document analysis assistant',
|
|
173
|
+
userMessage: 'Please analyze this document and extract key information',
|
|
174
|
+
messageHistory: [],
|
|
175
|
+
pdfDocuments: [pdfBuffer] // Uint8Array of PDF content
|
|
176
|
+
});
|
|
169
177
|
```
|
|
170
178
|
|
|
171
179
|
Both providers will:
|
|
@@ -175,7 +183,7 @@ Both providers will:
|
|
|
175
183
|
|
|
176
184
|
### Vision Processing
|
|
177
185
|
|
|
178
|
-
For providers that support vision tasks (OpenAI and
|
|
186
|
+
For providers that support vision tasks (OpenAI, Ollama, and Anthropic):
|
|
179
187
|
|
|
180
188
|
```typescript
|
|
181
189
|
// Using OpenAI's GPT-4 Vision
|
|
@@ -189,6 +197,12 @@ const analysis = await smartAi.ollamaProvider.vision({
|
|
|
189
197
|
image: imageBuffer,
|
|
190
198
|
prompt: 'Analyze this image in detail'
|
|
191
199
|
});
|
|
200
|
+
|
|
201
|
+
// Using Anthropic's Claude 3
|
|
202
|
+
const anthropicAnalysis = await smartAi.anthropicProvider.vision({
|
|
203
|
+
image: imageBuffer,
|
|
204
|
+
prompt: 'Please analyze this image and describe what you see'
|
|
205
|
+
});
|
|
192
206
|
```
|
|
193
207
|
|
|
194
208
|
## Error Handling
|
package/ts/00_commitinfo_data.ts
CHANGED
package/ts/provider.anthropic.ts
CHANGED
|
@@ -2,6 +2,9 @@ import * as plugins from './plugins.js';
|
|
|
2
2
|
import * as paths from './paths.js';
|
|
3
3
|
import { MultiModalModel } from './abstract.classes.multimodal.js';
|
|
4
4
|
import type { ChatOptions, ChatResponse, ChatMessage } from './abstract.classes.multimodal.js';
|
|
5
|
+
import type { ImageBlockParam, TextBlockParam } from '@anthropic-ai/sdk/resources/messages';
|
|
6
|
+
|
|
7
|
+
type ContentBlock = ImageBlockParam | TextBlockParam;
|
|
5
8
|
|
|
6
9
|
export interface IAnthropicProviderOptions {
|
|
7
10
|
anthropicToken: string;
|
|
@@ -132,7 +135,40 @@ export class AnthropicProvider extends MultiModalModel {
|
|
|
132
135
|
}
|
|
133
136
|
|
|
134
137
|
public async vision(optionsArg: { image: Buffer; prompt: string }): Promise<string> {
|
|
135
|
-
|
|
138
|
+
const base64Image = optionsArg.image.toString('base64');
|
|
139
|
+
|
|
140
|
+
const content: ContentBlock[] = [
|
|
141
|
+
{
|
|
142
|
+
type: 'text',
|
|
143
|
+
text: optionsArg.prompt
|
|
144
|
+
},
|
|
145
|
+
{
|
|
146
|
+
type: 'image',
|
|
147
|
+
source: {
|
|
148
|
+
type: 'base64',
|
|
149
|
+
media_type: 'image/jpeg',
|
|
150
|
+
data: base64Image
|
|
151
|
+
}
|
|
152
|
+
}
|
|
153
|
+
];
|
|
154
|
+
|
|
155
|
+
const result = await this.anthropicApiClient.messages.create({
|
|
156
|
+
model: 'claude-3-opus-20240229',
|
|
157
|
+
messages: [{
|
|
158
|
+
role: 'user',
|
|
159
|
+
content
|
|
160
|
+
}],
|
|
161
|
+
max_tokens: 1024
|
|
162
|
+
});
|
|
163
|
+
|
|
164
|
+
// Extract text content from the response
|
|
165
|
+
let message = '';
|
|
166
|
+
for (const block of result.content) {
|
|
167
|
+
if ('text' in block) {
|
|
168
|
+
message += block.text;
|
|
169
|
+
}
|
|
170
|
+
}
|
|
171
|
+
return message;
|
|
136
172
|
}
|
|
137
173
|
|
|
138
174
|
public async document(optionsArg: {
|
|
@@ -141,6 +177,64 @@ export class AnthropicProvider extends MultiModalModel {
|
|
|
141
177
|
pdfDocuments: Uint8Array[];
|
|
142
178
|
messageHistory: ChatMessage[];
|
|
143
179
|
}): Promise<{ message: any }> {
|
|
144
|
-
|
|
180
|
+
// Convert PDF documents to images using SmartPDF
|
|
181
|
+
const smartpdfInstance = new plugins.smartpdf.SmartPdf();
|
|
182
|
+
let documentImageBytesArray: Uint8Array[] = [];
|
|
183
|
+
|
|
184
|
+
for (const pdfDocument of optionsArg.pdfDocuments) {
|
|
185
|
+
const documentImageArray = await smartpdfInstance.convertPDFToPngBytes(pdfDocument);
|
|
186
|
+
documentImageBytesArray = documentImageBytesArray.concat(documentImageArray);
|
|
187
|
+
}
|
|
188
|
+
|
|
189
|
+
// Convert message history to Anthropic format
|
|
190
|
+
const messages = optionsArg.messageHistory.map(msg => ({
|
|
191
|
+
role: msg.role === 'assistant' ? 'assistant' as const : 'user' as const,
|
|
192
|
+
content: msg.content
|
|
193
|
+
}));
|
|
194
|
+
|
|
195
|
+
// Create content array with text and images
|
|
196
|
+
const content: ContentBlock[] = [
|
|
197
|
+
{
|
|
198
|
+
type: 'text',
|
|
199
|
+
text: optionsArg.userMessage
|
|
200
|
+
}
|
|
201
|
+
];
|
|
202
|
+
|
|
203
|
+
// Add each document page as an image
|
|
204
|
+
for (const imageBytes of documentImageBytesArray) {
|
|
205
|
+
content.push({
|
|
206
|
+
type: 'image',
|
|
207
|
+
source: {
|
|
208
|
+
type: 'base64',
|
|
209
|
+
media_type: 'image/jpeg',
|
|
210
|
+
data: Buffer.from(imageBytes).toString('base64')
|
|
211
|
+
}
|
|
212
|
+
});
|
|
213
|
+
}
|
|
214
|
+
|
|
215
|
+
const result = await this.anthropicApiClient.messages.create({
|
|
216
|
+
model: 'claude-3-opus-20240229',
|
|
217
|
+
system: optionsArg.systemMessage,
|
|
218
|
+
messages: [
|
|
219
|
+
...messages,
|
|
220
|
+
{ role: 'user', content }
|
|
221
|
+
],
|
|
222
|
+
max_tokens: 4096
|
|
223
|
+
});
|
|
224
|
+
|
|
225
|
+
// Extract text content from the response
|
|
226
|
+
let message = '';
|
|
227
|
+
for (const block of result.content) {
|
|
228
|
+
if ('text' in block) {
|
|
229
|
+
message += block.text;
|
|
230
|
+
}
|
|
231
|
+
}
|
|
232
|
+
|
|
233
|
+
return {
|
|
234
|
+
message: {
|
|
235
|
+
role: 'assistant',
|
|
236
|
+
content: message
|
|
237
|
+
}
|
|
238
|
+
};
|
|
145
239
|
}
|
|
146
240
|
}
|