roboto-js 1.4.50 → 1.5.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/cjs/index.cjs +77 -58
- package/dist/cjs/rbt_api.cjs +184 -144
- package/dist/cjs/rbt_file.cjs +71 -425
- package/dist/esm/index.js +3 -0
- package/dist/esm/rbt_api.js +18 -0
- package/dist/esm/rbt_file.js +61 -170
- package/package.json +3 -2
- package/src/index.js +3 -0
- package/src/rbt_api.js +21 -0
- package/src/rbt_file.js +65 -211
package/dist/esm/rbt_file.js
CHANGED
|
@@ -1,17 +1,16 @@
|
|
|
1
|
+
import { Upload } from 'tus-js-client';
|
|
1
2
|
import _ from 'lodash';
|
|
2
3
|
import EventEmitter from 'eventemitter3';
|
|
3
4
|
export default class RbtFile extends EventEmitter {
|
|
4
|
-
|
|
5
|
-
|
|
6
|
-
constructor(record, axiosInstance, localDb) {
|
|
5
|
+
constructor(record, axiosInstance) {
|
|
7
6
|
super(); // Call the constructor of EventEmitter
|
|
7
|
+
|
|
8
|
+
this.isRbtFile = true;
|
|
8
9
|
this.id = record.id;
|
|
9
10
|
this._axios = axiosInstance;
|
|
10
11
|
this._internalData = record;
|
|
11
12
|
this._data = record.data ? record.data : record.dataJson ? JSON.parse(record.dataJson) : {};
|
|
12
|
-
this._localDb = localDb;
|
|
13
13
|
this.progress = 0;
|
|
14
|
-
this.fileHeader = {};
|
|
15
14
|
}
|
|
16
15
|
get(path) {
|
|
17
16
|
return _.get(this._data, path);
|
|
@@ -24,6 +23,13 @@ export default class RbtFile extends EventEmitter {
|
|
|
24
23
|
...this._data
|
|
25
24
|
};
|
|
26
25
|
}
|
|
26
|
+
setData(data) {
|
|
27
|
+
// Use lodash's merge to deeply merge the incoming data into _data
|
|
28
|
+
_.merge(this._data, data);
|
|
29
|
+
return {
|
|
30
|
+
...this._data
|
|
31
|
+
}; // Return the updated _data as a new object
|
|
32
|
+
}
|
|
27
33
|
toRecord() {
|
|
28
34
|
return {
|
|
29
35
|
...this._internalData,
|
|
@@ -53,7 +59,7 @@ export default class RbtFile extends EventEmitter {
|
|
|
53
59
|
}
|
|
54
60
|
try {
|
|
55
61
|
const record = this.toRecord();
|
|
56
|
-
const response = await this._axios.
|
|
62
|
+
const response = await this._axios.delete(`/object_service/deleteObject/${this.id}`);
|
|
57
63
|
if (response.data.ok === false) {
|
|
58
64
|
throw new Error(response.data.message);
|
|
59
65
|
}
|
|
@@ -64,175 +70,60 @@ export default class RbtFile extends EventEmitter {
|
|
|
64
70
|
throw e; // Propagate the error
|
|
65
71
|
}
|
|
66
72
|
}
|
|
67
|
-
|
|
68
|
-
const
|
|
69
|
-
|
|
70
|
-
|
|
71
|
-
|
|
72
|
-
|
|
73
|
-
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
|
|
80
|
-
|
|
81
|
-
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
|
|
85
|
-
|
|
86
|
-
|
|
87
|
-
name: file.name,
|
|
88
|
-
size: file.size,
|
|
89
|
-
ext: this._extractFileExtension(file.name),
|
|
90
|
-
mimeType: file.type,
|
|
91
|
-
numFrames: Math.ceil(file.size / RbtFile.FRAME_SIZE)
|
|
92
|
-
};
|
|
93
|
-
while (offset < file.size) {
|
|
94
|
-
const chunk = file.slice(offset, offset + RbtFile.FRAME_SIZE);
|
|
95
|
-
await new Promise((resolve, reject) => {
|
|
96
|
-
const reader = new FileReader();
|
|
97
|
-
reader.onload = async e => {
|
|
98
|
-
const arrayBuffer = e.target.result;
|
|
99
|
-
const frameIndex = offset / RbtFile.FRAME_SIZE;
|
|
100
|
-
await this._storeChunkInIDB(arrayBuffer, frameIndex, this.fileHeader);
|
|
101
|
-
this.setProgress((frameIndex + 1) / this.fileHeader.numFrames);
|
|
102
|
-
resolve();
|
|
103
|
-
};
|
|
104
|
-
reader.onerror = () => reject(reader.error);
|
|
105
|
-
reader.readAsArrayBuffer(chunk);
|
|
106
|
-
});
|
|
107
|
-
offset += RbtFile.FRAME_SIZE;
|
|
108
|
-
}
|
|
109
|
-
}
|
|
110
|
-
_extractFileExtension(fileName) {
|
|
111
|
-
// Find the last dot in the filename
|
|
112
|
-
const lastDotIndex = fileName.lastIndexOf('.');
|
|
113
|
-
|
|
114
|
-
// No dot found, or the dot is the first character (hidden files)
|
|
115
|
-
if (lastDotIndex === -1 || lastDotIndex === 0) return '';
|
|
73
|
+
uploadFile(file) {
|
|
74
|
+
const upload = new Upload(file, {
|
|
75
|
+
endpoint: this._axios.defaults.baseURL + "/file_service/files",
|
|
76
|
+
retryDelays: [0, 1000, 3000, 5000],
|
|
77
|
+
metadata: {
|
|
78
|
+
rbtfileid: this.id,
|
|
79
|
+
filename: file.name,
|
|
80
|
+
filetype: file.type
|
|
81
|
+
},
|
|
82
|
+
onError: error => {
|
|
83
|
+
console.error("Failed because:", error);
|
|
84
|
+
this.emit('error', error);
|
|
85
|
+
},
|
|
86
|
+
onProgress: (bytesUploaded, bytesTotal) => {
|
|
87
|
+
console.log("File progress", bytesUploaded, bytesTotal);
|
|
88
|
+
const percentage = (bytesUploaded / bytesTotal * 100).toFixed(2);
|
|
89
|
+
this.setProgress(percentage / 100); // Update progress normalized between 0 and 1
|
|
90
|
+
},
|
|
91
|
+
onSuccess: () => {
|
|
92
|
+
console.log("File uploaded to:", upload.url);
|
|
116
93
|
|
|
117
|
-
|
|
118
|
-
|
|
119
|
-
|
|
120
|
-
|
|
121
|
-
|
|
122
|
-
|
|
123
|
-
|
|
124
|
-
|
|
125
|
-
|
|
126
|
-
|
|
127
|
-
const frameKey = `${this.id}_${frameIndex}`; // Unique key combining id and frameIndex
|
|
128
|
-
await tx.store.put({
|
|
129
|
-
key: frameKey,
|
|
130
|
-
chunk,
|
|
131
|
-
fileHeader
|
|
132
|
-
}, frameKey);
|
|
133
|
-
await tx.done;
|
|
134
|
-
}
|
|
135
|
-
async _readFrameFromIDB(frameIndex) {
|
|
136
|
-
const db = this._localDb;
|
|
137
|
-
if (!db) {
|
|
138
|
-
console.error('Database not initialized');
|
|
139
|
-
return null;
|
|
140
|
-
}
|
|
141
|
-
const frameKey = `${this.id}_${frameIndex}`; // Same key as used in _storeChunkInIDB
|
|
142
|
-
const tx = db.transaction('files', 'readonly');
|
|
143
|
-
const frame = await tx.store.get(frameKey);
|
|
144
|
-
return frame ? frame.chunk : null;
|
|
145
|
-
}
|
|
146
|
-
async uploadFile() {
|
|
147
|
-
if (!this.fileHeader || !this.fileHeader.numFrames) {
|
|
148
|
-
throw new Error("File not ready for upload.");
|
|
149
|
-
}
|
|
150
|
-
this.fileRecord = this.toRecord();
|
|
151
|
-
let frameRes;
|
|
152
|
-
for (let frameIndex = 0; frameIndex < this.fileHeader.numFrames; frameIndex++) {
|
|
153
|
-
try {
|
|
154
|
-
const frameData = await this._readFrameFromIDB(frameIndex);
|
|
155
|
-
if (!frameData) {
|
|
156
|
-
throw new Error(`Failed to read frame ${frameIndex}`);
|
|
157
|
-
}
|
|
158
|
-
frameRes = await this._uploadFrameToServer(frameData, frameIndex, this.fileHeader, this.fileRecord);
|
|
159
|
-
|
|
160
|
-
// Update progress after each frame is uploaded
|
|
161
|
-
this.setProgress((frameIndex + 1) / this.fileHeader.numFrames);
|
|
162
|
-
} catch (error) {
|
|
163
|
-
console.error(`Error uploading frame ${frameIndex}:`, error);
|
|
164
|
-
throw error; // Stop the upload process if an error occurs
|
|
165
|
-
}
|
|
166
|
-
}
|
|
167
|
-
|
|
168
|
-
//
|
|
169
|
-
// DONE UPLOADING, PROCESSING
|
|
170
|
-
//
|
|
171
|
-
|
|
172
|
-
try {
|
|
173
|
-
const statusRes = await this._checkUploadStatusUntilDone(this.fileRecord);
|
|
174
|
-
this.set('remoteSrc', statusRes.remoteSrc);
|
|
175
|
-
this.set('progress', 1);
|
|
176
|
-
await this.save();
|
|
177
|
-
} catch (error) {
|
|
178
|
-
console.error('Error during upload status check:', error);
|
|
179
|
-
// Handle error, possibly update UI or retry logic
|
|
180
|
-
}
|
|
181
|
-
}
|
|
182
|
-
async _uploadFrameToServer(frameData, frameIndex, fileHeader, fileRecord) {
|
|
183
|
-
const base64Data = this.convertToBase64(frameData);
|
|
184
|
-
const payload = ["@filekit.ffs_server_receiveAndStore_toS3_flow", {
|
|
185
|
-
"fileObject": fileRecord,
|
|
186
|
-
// using fileRecord from newFile.toRecord()
|
|
187
|
-
"fileHeader": fileHeader,
|
|
188
|
-
"frameIndex": frameIndex,
|
|
189
|
-
"frameData": `rpcBase64:${base64Data}`,
|
|
190
|
-
"timeout": 86400000
|
|
191
|
-
}];
|
|
192
|
-
|
|
193
|
-
//console.log(payload);
|
|
194
|
-
|
|
195
|
-
const response = await this._axios.post('/file_service/ffs_runFlow', payload);
|
|
196
|
-
if (!response || response.status != 200) {
|
|
197
|
-
throw new Error('Error uploading frame to server');
|
|
198
|
-
}
|
|
199
|
-
return response;
|
|
200
|
-
}
|
|
201
|
-
async _checkUploadStatusUntilDone(fileRecord) {
|
|
202
|
-
return new Promise((resolve, reject) => {
|
|
203
|
-
const intervalId = setInterval(async () => {
|
|
204
|
-
try {
|
|
205
|
-
const statusRes = await this._uploadGetStatusUpdate(fileRecord);
|
|
206
|
-
if (statusRes.data && statusRes.data.fileHeader && statusRes.data.fileHeader.status === 'DONE') {
|
|
207
|
-
clearInterval(intervalId); // Stop the interval
|
|
208
|
-
resolve(statusRes.data.fileHeader); // Resolve the promise with the final status
|
|
209
|
-
} else {
|
|
210
|
-
// Optionally update progress or handle other status cases here
|
|
211
|
-
//this.emit('progress', statusRes.progress); // Emit progress updates if available
|
|
94
|
+
// Remove the `options.endpoint` from the start of the `url`
|
|
95
|
+
const m1 = upload.url.match(/\/([^\/]+)$/);
|
|
96
|
+
const remoteId = m1 ? m1[1] : null;
|
|
97
|
+
this.setData({
|
|
98
|
+
'remoteId': remoteId,
|
|
99
|
+
'sourceFile': {
|
|
100
|
+
name: upload.file.name,
|
|
101
|
+
size: upload.file.size,
|
|
102
|
+
type: upload.file.type,
|
|
103
|
+
lastModified: upload.file.lastModified
|
|
212
104
|
}
|
|
213
|
-
}
|
|
214
|
-
|
|
215
|
-
|
|
216
|
-
}
|
|
217
|
-
}
|
|
105
|
+
});
|
|
106
|
+
this.save().then(() => {
|
|
107
|
+
this.emit('success', upload.url);
|
|
108
|
+
});
|
|
109
|
+
}
|
|
218
110
|
});
|
|
219
|
-
}
|
|
220
|
-
async _uploadGetStatusUpdate(fileRecord) {
|
|
221
|
-
const payload = ["@filekit.ffs_server_receiveAndStore_toS3_flow", {
|
|
222
|
-
"fileObject": fileRecord,
|
|
223
|
-
// using fileRecord from newFile.toRecord()
|
|
224
|
-
"getStatusUpdate": true
|
|
225
|
-
}];
|
|
226
111
|
|
|
227
|
-
//
|
|
112
|
+
// Check if there are any previous uploads to continue.
|
|
113
|
+
// upload.findPreviousUploads().then(function (previousUploads) {
|
|
114
|
+
// // Found previous uploads so we select the first one.
|
|
115
|
+
// if (previousUploads.length) {
|
|
116
|
+
// upload.resumeFromPreviousUpload(previousUploads[0])
|
|
117
|
+
// }
|
|
228
118
|
|
|
229
|
-
|
|
230
|
-
|
|
231
|
-
|
|
232
|
-
|
|
233
|
-
return response;
|
|
119
|
+
// // Start the upload
|
|
120
|
+
// upload.start();
|
|
121
|
+
// })
|
|
122
|
+
upload.start();
|
|
234
123
|
}
|
|
235
|
-
|
|
236
|
-
|
|
124
|
+
setProgress(newProgress) {
|
|
125
|
+
this.progress = newProgress;
|
|
126
|
+
console.log(`Progress: ${this.progress * 100}%`);
|
|
127
|
+
this.emit('progress', this.progress); // Emit a progress event
|
|
237
128
|
}
|
|
238
129
|
}
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "roboto-js",
|
|
3
|
-
"version": "1.
|
|
3
|
+
"version": "1.5.0",
|
|
4
4
|
"type": "module",
|
|
5
5
|
"description": "",
|
|
6
6
|
"main": "dist/cjs/index.cjs",
|
|
@@ -25,7 +25,8 @@
|
|
|
25
25
|
"crypto-js": "^4.1.1",
|
|
26
26
|
"eventemitter3": "^5.0.1",
|
|
27
27
|
"idb": "^8.0.0",
|
|
28
|
-
"lodash": "^4.17.21"
|
|
28
|
+
"lodash": "^4.17.21",
|
|
29
|
+
"tus-js-client": "^4.2.3"
|
|
29
30
|
},
|
|
30
31
|
"devDependencies": {
|
|
31
32
|
"@babel/cli": "^7.23.9",
|
package/src/index.js
CHANGED
package/src/rbt_api.js
CHANGED
|
@@ -403,6 +403,27 @@ export default class RbtApi {
|
|
|
403
403
|
|
|
404
404
|
}
|
|
405
405
|
|
|
406
|
+
async loadFiles(ids) {
|
|
407
|
+
try {
|
|
408
|
+
// Use the bulk load method with the provided IDs
|
|
409
|
+
const responses = await this.load('<@filekit.file>', ids);
|
|
410
|
+
|
|
411
|
+
if (!responses || !Array.isArray(responses)) {
|
|
412
|
+
return []; // Return an empty array if no responses or invalid data
|
|
413
|
+
}
|
|
414
|
+
|
|
415
|
+
// Map over the responses to create RbtFile instances
|
|
416
|
+
return responses.map(response => {
|
|
417
|
+
const record = response.toRecord();
|
|
418
|
+
return new RbtFile(record, this.axios, this.localDb);
|
|
419
|
+
});
|
|
420
|
+
} catch (e) {
|
|
421
|
+
this._handleError(e); // Handle errors (log or process as needed)
|
|
422
|
+
return []; // Return an empty array on failure
|
|
423
|
+
}
|
|
424
|
+
}
|
|
425
|
+
|
|
426
|
+
|
|
406
427
|
|
|
407
428
|
/**
|
|
408
429
|
* Creates a new object of the given type.
|
package/src/rbt_file.js
CHANGED
|
@@ -1,22 +1,18 @@
|
|
|
1
|
-
|
|
1
|
+
import { Upload } from 'tus-js-client';
|
|
2
2
|
import _ from 'lodash';
|
|
3
3
|
import EventEmitter from 'eventemitter3';
|
|
4
4
|
|
|
5
|
-
export default class RbtFile extends EventEmitter{
|
|
6
|
-
|
|
7
|
-
static FRAME_SIZE = 1024 * 1024; // 1MB, for example
|
|
8
|
-
|
|
9
|
-
constructor(record, axiosInstance, localDb) {
|
|
5
|
+
export default class RbtFile extends EventEmitter {
|
|
10
6
|
|
|
7
|
+
constructor(record, axiosInstance) {
|
|
11
8
|
super(); // Call the constructor of EventEmitter
|
|
9
|
+
|
|
10
|
+
this.isRbtFile = true;
|
|
12
11
|
this.id = record.id;
|
|
13
12
|
this._axios = axiosInstance;
|
|
14
13
|
this._internalData = record;
|
|
15
14
|
this._data = record.data ? record.data : (record.dataJson ? JSON.parse(record.dataJson) : {});
|
|
16
|
-
this._localDb = localDb;
|
|
17
15
|
this.progress = 0;
|
|
18
|
-
|
|
19
|
-
this.fileHeader = {};
|
|
20
16
|
}
|
|
21
17
|
|
|
22
18
|
get(path) {
|
|
@@ -31,6 +27,12 @@ export default class RbtFile extends EventEmitter{
|
|
|
31
27
|
return { ...this._data };
|
|
32
28
|
}
|
|
33
29
|
|
|
30
|
+
setData(data) {
|
|
31
|
+
// Use lodash's merge to deeply merge the incoming data into _data
|
|
32
|
+
_.merge(this._data, data);
|
|
33
|
+
return { ...this._data }; // Return the updated _data as a new object
|
|
34
|
+
}
|
|
35
|
+
|
|
34
36
|
toRecord() {
|
|
35
37
|
return {
|
|
36
38
|
...this._internalData,
|
|
@@ -52,7 +54,6 @@ export default class RbtFile extends EventEmitter{
|
|
|
52
54
|
}
|
|
53
55
|
|
|
54
56
|
this._internalData = response.data;
|
|
55
|
-
|
|
56
57
|
return this;
|
|
57
58
|
|
|
58
59
|
} catch (e) {
|
|
@@ -68,7 +69,7 @@ export default class RbtFile extends EventEmitter{
|
|
|
68
69
|
|
|
69
70
|
try {
|
|
70
71
|
const record = this.toRecord();
|
|
71
|
-
const response = await this._axios.
|
|
72
|
+
const response = await this._axios.delete(`/object_service/deleteObject/${this.id}`);
|
|
72
73
|
|
|
73
74
|
if (response.data.ok === false) {
|
|
74
75
|
throw new Error(response.data.message);
|
|
@@ -83,210 +84,63 @@ export default class RbtFile extends EventEmitter{
|
|
|
83
84
|
}
|
|
84
85
|
}
|
|
85
86
|
|
|
86
|
-
|
|
87
|
-
|
|
88
|
-
const
|
|
89
|
-
|
|
90
|
-
|
|
91
|
-
|
|
92
|
-
|
|
93
|
-
|
|
94
|
-
|
|
95
|
-
|
|
96
|
-
|
|
97
|
-
|
|
98
|
-
|
|
99
|
-
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
|
|
103
|
-
|
|
104
|
-
|
|
105
|
-
|
|
106
|
-
|
|
107
|
-
|
|
108
|
-
|
|
109
|
-
|
|
110
|
-
|
|
111
|
-
|
|
112
|
-
|
|
113
|
-
|
|
114
|
-
|
|
115
|
-
|
|
116
|
-
|
|
117
|
-
|
|
118
|
-
|
|
119
|
-
|
|
120
|
-
|
|
121
|
-
|
|
122
|
-
|
|
123
|
-
|
|
124
|
-
resolve();
|
|
125
|
-
};
|
|
126
|
-
reader.onerror = () => reject(reader.error);
|
|
127
|
-
reader.readAsArrayBuffer(chunk);
|
|
128
|
-
});
|
|
129
|
-
offset += RbtFile.FRAME_SIZE;
|
|
130
|
-
}
|
|
131
|
-
}
|
|
132
|
-
|
|
133
|
-
_extractFileExtension(fileName) {
|
|
134
|
-
// Find the last dot in the filename
|
|
135
|
-
const lastDotIndex = fileName.lastIndexOf('.');
|
|
136
|
-
|
|
137
|
-
// No dot found, or the dot is the first character (hidden files)
|
|
138
|
-
if (lastDotIndex === -1 || lastDotIndex === 0) return '';
|
|
139
|
-
|
|
140
|
-
// Extract the extension
|
|
141
|
-
return fileName.substring(lastDotIndex + 1);
|
|
142
|
-
}
|
|
143
|
-
|
|
144
|
-
async _storeChunkInIDB(chunk, frameIndex, fileHeader) {
|
|
145
|
-
const db = this._localDb;
|
|
146
|
-
if (!db) {
|
|
147
|
-
console.error('Database not initialized');
|
|
148
|
-
return;
|
|
149
|
-
}
|
|
150
|
-
const tx = db.transaction('files', 'readwrite');
|
|
151
|
-
const frameKey = `${this.id}_${frameIndex}`; // Unique key combining id and frameIndex
|
|
152
|
-
await tx.store.put({ key: frameKey, chunk, fileHeader }, frameKey);
|
|
153
|
-
await tx.done;
|
|
154
|
-
}
|
|
155
|
-
|
|
156
|
-
async _readFrameFromIDB(frameIndex) {
|
|
157
|
-
const db = this._localDb;
|
|
158
|
-
if (!db) {
|
|
159
|
-
console.error('Database not initialized');
|
|
160
|
-
return null;
|
|
161
|
-
}
|
|
162
|
-
const frameKey = `${this.id}_${frameIndex}`; // Same key as used in _storeChunkInIDB
|
|
163
|
-
const tx = db.transaction('files', 'readonly');
|
|
164
|
-
const frame = await tx.store.get(frameKey);
|
|
165
|
-
|
|
166
|
-
return frame ? frame.chunk : null;
|
|
167
|
-
}
|
|
168
|
-
|
|
169
|
-
async uploadFile() {
|
|
170
|
-
if (!this.fileHeader || !this.fileHeader.numFrames) {
|
|
171
|
-
throw new Error("File not ready for upload.");
|
|
172
|
-
}
|
|
173
|
-
|
|
174
|
-
this.fileRecord = this.toRecord();
|
|
175
|
-
|
|
176
|
-
let frameRes;
|
|
177
|
-
|
|
178
|
-
for (let frameIndex = 0; frameIndex < this.fileHeader.numFrames; frameIndex++) {
|
|
179
|
-
try {
|
|
180
|
-
const frameData = await this._readFrameFromIDB(frameIndex);
|
|
181
|
-
if (!frameData) {
|
|
182
|
-
throw new Error(`Failed to read frame ${frameIndex}`);
|
|
183
|
-
}
|
|
184
|
-
|
|
185
|
-
frameRes = await this._uploadFrameToServer(frameData, frameIndex, this.fileHeader, this.fileRecord);
|
|
186
|
-
|
|
187
|
-
// Update progress after each frame is uploaded
|
|
188
|
-
this.setProgress((frameIndex + 1) / this.fileHeader.numFrames);
|
|
189
|
-
} catch (error) {
|
|
190
|
-
console.error(`Error uploading frame ${frameIndex}:`, error);
|
|
191
|
-
throw error; // Stop the upload process if an error occurs
|
|
192
|
-
}
|
|
193
|
-
}
|
|
194
|
-
|
|
195
|
-
|
|
196
|
-
//
|
|
197
|
-
// DONE UPLOADING, PROCESSING
|
|
198
|
-
//
|
|
199
|
-
|
|
200
|
-
try {
|
|
201
|
-
const statusRes = await this._checkUploadStatusUntilDone(this.fileRecord);
|
|
202
|
-
|
|
203
|
-
this.set('remoteSrc', statusRes.remoteSrc);
|
|
204
|
-
this.set('progress', 1);
|
|
205
|
-
await this.save();
|
|
206
|
-
} catch (error) {
|
|
207
|
-
console.error('Error during upload status check:', error);
|
|
208
|
-
// Handle error, possibly update UI or retry logic
|
|
209
|
-
}
|
|
210
|
-
|
|
211
|
-
}
|
|
212
|
-
|
|
213
|
-
async _uploadFrameToServer(frameData, frameIndex, fileHeader, fileRecord) {
|
|
214
|
-
|
|
215
|
-
const base64Data = this.convertToBase64(frameData);
|
|
216
|
-
const payload = [
|
|
217
|
-
"@filekit.ffs_server_receiveAndStore_toS3_flow",
|
|
218
|
-
{
|
|
219
|
-
"fileObject": fileRecord, // using fileRecord from newFile.toRecord()
|
|
220
|
-
"fileHeader": fileHeader,
|
|
221
|
-
"frameIndex": frameIndex,
|
|
222
|
-
"frameData": `rpcBase64:${base64Data}`,
|
|
223
|
-
"timeout": 86400000
|
|
224
|
-
}
|
|
225
|
-
];
|
|
226
|
-
|
|
227
|
-
//console.log(payload);
|
|
228
|
-
|
|
229
|
-
const response = await this._axios.post('/file_service/ffs_runFlow', payload);
|
|
230
|
-
|
|
231
|
-
if (!response || response.status != 200) {
|
|
232
|
-
throw new Error('Error uploading frame to server');
|
|
233
|
-
}
|
|
234
|
-
|
|
235
|
-
return response;
|
|
236
|
-
}
|
|
237
|
-
|
|
238
|
-
|
|
239
|
-
async _checkUploadStatusUntilDone(fileRecord) {
|
|
240
|
-
|
|
241
|
-
return new Promise((resolve, reject) => {
|
|
242
|
-
const intervalId = setInterval(async () => {
|
|
243
|
-
try {
|
|
244
|
-
const statusRes = await this._uploadGetStatusUpdate(fileRecord);
|
|
245
|
-
if (statusRes.data && statusRes.data.fileHeader && statusRes.data.fileHeader.status === 'DONE') {
|
|
246
|
-
clearInterval(intervalId); // Stop the interval
|
|
247
|
-
resolve(statusRes.data.fileHeader); // Resolve the promise with the final status
|
|
248
|
-
} else {
|
|
249
|
-
// Optionally update progress or handle other status cases here
|
|
250
|
-
//this.emit('progress', statusRes.progress); // Emit progress updates if available
|
|
87
|
+
uploadFile(file) {
|
|
88
|
+
|
|
89
|
+
const upload = new Upload(file, {
|
|
90
|
+
endpoint: this._axios.defaults.baseURL+"/file_service/files",
|
|
91
|
+
retryDelays: [0, 1000, 3000, 5000],
|
|
92
|
+
metadata: {
|
|
93
|
+
rbtfileid: this.id,
|
|
94
|
+
filename: file.name,
|
|
95
|
+
filetype: file.type,
|
|
96
|
+
},
|
|
97
|
+
onError: error => {
|
|
98
|
+
console.error("Failed because:", error);
|
|
99
|
+
this.emit('error', error);
|
|
100
|
+
},
|
|
101
|
+
onProgress: (bytesUploaded, bytesTotal) => {
|
|
102
|
+
console.log("File progress", bytesUploaded, bytesTotal);
|
|
103
|
+
const percentage = (bytesUploaded / bytesTotal * 100).toFixed(2);
|
|
104
|
+
this.setProgress(percentage / 100); // Update progress normalized between 0 and 1
|
|
105
|
+
},
|
|
106
|
+
onSuccess: () => {
|
|
107
|
+
console.log("File uploaded to:", upload.url);
|
|
108
|
+
|
|
109
|
+
// Remove the `options.endpoint` from the start of the `url`
|
|
110
|
+
const m1 = upload.url.match(/\/([^\/]+)$/);
|
|
111
|
+
const remoteId = m1 ? m1[1] : null;
|
|
112
|
+
|
|
113
|
+
this.setData({
|
|
114
|
+
'remoteId': remoteId,
|
|
115
|
+
'sourceFile': {
|
|
116
|
+
name: upload.file.name,
|
|
117
|
+
size: upload.file.size,
|
|
118
|
+
type: upload.file.type,
|
|
119
|
+
lastModified: upload.file.lastModified,
|
|
120
|
+
}
|
|
121
|
+
});
|
|
122
|
+
this.save().then(() => {
|
|
123
|
+
this.emit('success', upload.url);
|
|
124
|
+
});
|
|
251
125
|
}
|
|
252
|
-
|
|
253
|
-
clearInterval(intervalId); // Stop the interval on error
|
|
254
|
-
reject(error); // Reject the promise if there's an error
|
|
255
|
-
}
|
|
256
|
-
}, 2000); // Check every 2 seconds
|
|
257
|
-
});
|
|
258
|
-
|
|
259
|
-
}
|
|
260
|
-
|
|
261
|
-
|
|
262
|
-
|
|
263
|
-
async _uploadGetStatusUpdate(fileRecord){
|
|
264
|
-
|
|
265
|
-
const payload = [
|
|
266
|
-
"@filekit.ffs_server_receiveAndStore_toS3_flow",
|
|
267
|
-
{
|
|
268
|
-
"fileObject": fileRecord, // using fileRecord from newFile.toRecord()
|
|
269
|
-
"getStatusUpdate": true
|
|
270
|
-
}
|
|
271
|
-
];
|
|
272
|
-
|
|
273
|
-
//console.log(payload);
|
|
274
|
-
|
|
275
|
-
const response = await this._axios.post('/file_service/ffs_runFlow', payload);
|
|
276
|
-
|
|
277
|
-
if (!response || response.status != 200) {
|
|
278
|
-
throw new Error('Error getting upload status');
|
|
279
|
-
}
|
|
280
|
-
|
|
281
|
-
return response;
|
|
126
|
+
});
|
|
282
127
|
|
|
128
|
+
// Check if there are any previous uploads to continue.
|
|
129
|
+
// upload.findPreviousUploads().then(function (previousUploads) {
|
|
130
|
+
// // Found previous uploads so we select the first one.
|
|
131
|
+
// if (previousUploads.length) {
|
|
132
|
+
// upload.resumeFromPreviousUpload(previousUploads[0])
|
|
133
|
+
// }
|
|
283
134
|
|
|
135
|
+
// // Start the upload
|
|
136
|
+
// upload.start();
|
|
137
|
+
// })
|
|
138
|
+
upload.start();
|
|
284
139
|
}
|
|
285
140
|
|
|
286
|
-
|
|
287
|
-
|
|
288
|
-
|
|
289
|
-
|
|
290
|
-
);
|
|
141
|
+
setProgress(newProgress) {
|
|
142
|
+
this.progress = newProgress;
|
|
143
|
+
console.log(`Progress: ${this.progress * 100}%`);
|
|
144
|
+
this.emit('progress', this.progress); // Emit a progress event
|
|
291
145
|
}
|
|
292
146
|
}
|