@seedprotocol/sdk 0.1.111 → 0.1.112
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/{index-BM9Ymz63.js → index-CwNnbNh8.js} +2 -2
- package/dist/{index-BM9Ymz63.js.map → index-CwNnbNh8.js.map} +1 -1
- package/dist/{index-CkDKrj91.js → index-OiUariqS.js} +9 -6
- package/dist/{index-CkDKrj91.js.map → index-OiUariqS.js.map} +1 -1
- package/dist/main.js +1 -1
- package/dist/{seed.schema.config-DfASQOur.js → seed.schema.config-CY5ZbFqn.js} +2 -2
- package/dist/{seed.schema.config-DfASQOur.js.map → seed.schema.config-CY5ZbFqn.js.map} +1 -1
- package/dist/src/actors.ts +33 -273
- package/dist/src/getPublishUploads.ts +6 -2
- package/dist/src/index.ts +28 -43
- package/dist/types/src/browser/db/read/getPublishUploads.d.ts.map +1 -1
- package/package.json +1 -1
package/dist/main.js
CHANGED
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
export { j as ImageSrc, I as Item, k as ItemProperty, J as Json, L as List, M as Model, P as Property, R as Relation, T as Text, t as client, v as getCorrectId, s as getGlobalService, n as useCreateItem, p as useDeleteItem, q as useGlobalServiceStatus, l as useItem, m as useItemProperties, o as useItemProperty, u as useItems, r as useServices, w as withSeed } from './index-
|
|
1
|
+
export { j as ImageSrc, I as Item, k as ItemProperty, J as Json, L as List, M as Model, P as Property, R as Relation, T as Text, t as client, v as getCorrectId, s as getGlobalService, n as useCreateItem, p as useDeleteItem, q as useGlobalServiceStatus, l as useItem, m as useItemProperties, o as useItemProperty, u as useItems, r as useServices, w as withSeed } from './index-OiUariqS.js';
|
|
2
2
|
import 'immer';
|
|
3
3
|
import 'reflect-metadata';
|
|
4
4
|
import './constants-CEnkzyny.js';
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
import { _ as __decorate, a as __metadata } from './constants-CEnkzyny.js';
|
|
2
2
|
import 'drizzle-orm';
|
|
3
|
-
import { T as Text, M as Model } from './index-
|
|
3
|
+
import { T as Text, M as Model } from './index-OiUariqS.js';
|
|
4
4
|
import 'react';
|
|
5
5
|
import 'reflect-metadata';
|
|
6
6
|
import 'xstate';
|
|
@@ -76,4 +76,4 @@ const models = {
|
|
|
76
76
|
};
|
|
77
77
|
|
|
78
78
|
export { models };
|
|
79
|
-
//# sourceMappingURL=seed.schema.config-
|
|
79
|
+
//# sourceMappingURL=seed.schema.config-CY5ZbFqn.js.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"seed.schema.config-
|
|
1
|
+
{"version":3,"file":"seed.schema.config-CY5ZbFqn.js","sources":["../../src/shared/configs/seed.schema.config.ts"],"sourcesContent":["import { Model, Text } from '@/browser'\n\n@Model\nclass Seed {\n @Text() uid!: string\n @Text() type!: string\n}\n\n@Model\nclass Version {\n @Text() seedUid!: string\n @Text() note!: string\n}\n\n@Model\nclass Metadata {\n @Text() key!: string\n @Text() value!: string\n}\n\nexport const models = {\n Seed,\n Version,\n Metadata,\n}\n"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAGA,IAAM,IAAI,GAAV,MAAM,IAAI,CAAA;CAGT;AAFS,UAAA,CAAA;AAAP,IAAA,IAAI,EAAE;;AAAa,CAAA,EAAA,IAAA,CAAA,SAAA,EAAA,KAAA,EAAA,KAAA,CAAA,CAAA;AACZ,UAAA,CAAA;AAAP,IAAA,IAAI,EAAE;;AAAc,CAAA,EAAA,IAAA,CAAA,SAAA,EAAA,MAAA,EAAA,KAAA,CAAA,CAAA;AAFjB,IAAI,GAAA,UAAA,CAAA;IADT;AACK,CAAA,EAAA,IAAI,CAGT;AAGD,IAAM,OAAO,GAAb,MAAM,OAAO,CAAA;CAGZ;AAFS,UAAA,CAAA;AAAP,IAAA,IAAI,EAAE;;AAAiB,CAAA,EAAA,OAAA,CAAA,SAAA,EAAA,SAAA,EAAA,KAAA,CAAA,CAAA;AAChB,UAAA,CAAA;AAAP,IAAA,IAAI,EAAE;;AAAc,CAAA,EAAA,OAAA,CAAA,SAAA,EAAA,MAAA,EAAA,KAAA,CAAA,CAAA;AAFjB,OAAO,GAAA,UAAA,CAAA;IADZ;AACK,CAAA,EAAA,OAAO,CAGZ;AAGD,IAAM,QAAQ,GAAd,MAAM,QAAQ,CAAA;CAGb;AAFS,UAAA,CAAA;AAAP,IAAA,IAAI,EAAE;;AAAa,CAAA,EAAA,QAAA,CAAA,SAAA,EAAA,KAAA,EAAA,KAAA,CAAA,CAAA;AACZ,UAAA,CAAA;AAAP,IAAA,IAAI,EAAE;;AAAe,CAAA,EAAA,QAAA,CAAA,SAAA,EAAA,OAAA,EAAA,KAAA,CAAA,CAAA;AAFlB,QAAQ,GAAA,UAAA,CAAA;IADb;AACK,CAAA,EAAA,QAAQ,CAGb;AAEY,MAAA,MAAM,GAAG;IACpB,IAAI;IACJ,OAAO;IACP,QAAQ;;;;;"}
|
package/dist/src/actors.ts
CHANGED
|
@@ -1,291 +1,51 @@
|
|
|
1
1
|
import { EventObject, fromCallback } from 'xstate'
|
|
2
|
-
import {
|
|
3
|
-
import { ARWEAVE_HOST } from '@/browser/services/internal/constants'
|
|
2
|
+
import { downloadMachine } from '@/browser/schema/file/download/index'
|
|
4
3
|
import { GET_FILES_METADATA } from '@/browser/schema/file/queries'
|
|
5
|
-
import { getArweave } from '../../../helpers/arweave'
|
|
6
|
-
import { fs } from '@zenfs/core'
|
|
7
|
-
import {
|
|
8
|
-
getDataTypeFromString,
|
|
9
|
-
getMimeType,
|
|
10
|
-
identifyString,
|
|
11
|
-
} from '@/shared/helpers'
|
|
12
|
-
import { appState } from 'src/shared/seedSchema'
|
|
13
|
-
import { eq } from 'drizzle-orm'
|
|
14
4
|
import { easClient, queryClient } from '@/browser/helpers'
|
|
15
|
-
import debug from 'debug'
|
|
16
|
-
import { getAppDb } from '@/browser/db/sqlWasmClient'
|
|
17
5
|
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
const
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
// Add more MIME types and file extensions as needed
|
|
34
|
-
}
|
|
35
|
-
|
|
36
|
-
export const fetchAllMetadataRecords = fromCallback<
|
|
37
|
-
EventObject,
|
|
38
|
-
typeof fetchAllFilesMachine
|
|
39
|
-
>(({ sendBack, receive, input: { context, event } }) => {
|
|
40
|
-
const { addresses } = context
|
|
41
|
-
|
|
42
|
-
const _fetchAllMetadataRecords = async () => {
|
|
43
|
-
const { filesMetadata } = await queryClient.fetchQuery({
|
|
44
|
-
queryKey: ['getFilesMetadata', ...addresses],
|
|
45
|
-
queryFn: async () =>
|
|
46
|
-
easClient.request(GET_FILES_METADATA, {
|
|
47
|
-
where: {
|
|
48
|
-
attester: {
|
|
49
|
-
in: addresses,
|
|
50
|
-
},
|
|
51
|
-
schema: {
|
|
52
|
-
is: {
|
|
53
|
-
id: {
|
|
54
|
-
equals:
|
|
55
|
-
'0x55fdefb36fcbbaebeb7d6b41dc3a1a9666e4e42154267c889de064faa7ede517',
|
|
56
|
-
},
|
|
6
|
+
export const fetchMetadata = fromCallback<EventObject, typeof downloadMachine>(
|
|
7
|
+
({ sendBack, input: { context } }) => {
|
|
8
|
+
const { addresses } = context
|
|
9
|
+
|
|
10
|
+
const fetchMetadata = async () => {
|
|
11
|
+
const metadataRecords = await queryClient.fetchQuery({
|
|
12
|
+
queryKey: ['getFilesMetadata', ...addresses],
|
|
13
|
+
queryFn: async () =>
|
|
14
|
+
easClient.request(GET_FILES_METADATA, {
|
|
15
|
+
where: {
|
|
16
|
+
attester: {
|
|
17
|
+
in: addresses,
|
|
18
|
+
},
|
|
19
|
+
decodedDataJson: {
|
|
20
|
+
contains: 'transactionId',
|
|
57
21
|
},
|
|
58
22
|
},
|
|
59
|
-
},
|
|
60
|
-
|
|
61
|
-
})
|
|
62
|
-
|
|
63
|
-
return filesMetadata
|
|
64
|
-
}
|
|
65
|
-
|
|
66
|
-
_fetchAllMetadataRecords().then((filesMetadata) => {
|
|
67
|
-
sendBack({ type: 'fetchingAllMetadataRecordsSuccess', filesMetadata })
|
|
68
|
-
})
|
|
69
|
-
|
|
70
|
-
return () => {}
|
|
71
|
-
})
|
|
23
|
+
}),
|
|
24
|
+
})
|
|
72
25
|
|
|
73
|
-
|
|
74
|
-
EventObject,
|
|
75
|
-
typeof fetchAllFilesMachine
|
|
76
|
-
>(({ sendBack, input: { context } }) => {
|
|
77
|
-
const { filesMetadata, addresses } = context
|
|
78
|
-
|
|
79
|
-
const _fetchAllBinaryData = async () => {
|
|
80
|
-
if (!(await fs.promises.exists('/files'))) {
|
|
81
|
-
await fs.promises.mkdir('/files', { recursive: true })
|
|
82
|
-
}
|
|
83
|
-
|
|
84
|
-
if (!(await fs.promises.exists('/files/html'))) {
|
|
85
|
-
await fs.promises.mkdir('/files/html', { recursive: true })
|
|
86
|
-
}
|
|
87
|
-
|
|
88
|
-
if (!(await fs.promises.exists('/files/json'))) {
|
|
89
|
-
await fs.promises.mkdir('/files/json', { recursive: true })
|
|
90
|
-
}
|
|
91
|
-
|
|
92
|
-
if (!(await fs.promises.exists('/files/images'))) {
|
|
93
|
-
await fs.promises.mkdir('/files/images', { recursive: true })
|
|
26
|
+
return metadataRecords
|
|
94
27
|
}
|
|
95
28
|
|
|
96
|
-
|
|
97
|
-
|
|
98
|
-
|
|
99
|
-
logger('[fetchAll/actors] [fetchAllBinaryData] seedDb not available')
|
|
100
|
-
return []
|
|
101
|
-
}
|
|
102
|
-
|
|
103
|
-
for (const fileMetadata of filesMetadata) {
|
|
104
|
-
const json = JSON.parse(fileMetadata.decodedDataJson)
|
|
105
|
-
const transactionId = json[0].value.value
|
|
106
|
-
|
|
107
|
-
const excludedTransactionsQuery = await appDb
|
|
108
|
-
.select()
|
|
109
|
-
.from(appState)
|
|
110
|
-
.where(eq(appState.key, 'excludedTransactions'))
|
|
111
|
-
|
|
112
|
-
let excludedTransactions = new Set<string>()
|
|
113
|
-
|
|
114
|
-
if (excludedTransactionsQuery && excludedTransactionsQuery.length === 1) {
|
|
115
|
-
const valueString = excludedTransactionsQuery[0].value
|
|
116
|
-
if (valueString) {
|
|
117
|
-
const excludedTransactionsArray = JSON.parse(valueString)
|
|
118
|
-
excludedTransactions = new Set(excludedTransactionsArray)
|
|
119
|
-
}
|
|
120
|
-
}
|
|
121
|
-
|
|
122
|
-
if (excludedTransactions.has(transactionId)) {
|
|
123
|
-
continue
|
|
124
|
-
}
|
|
125
|
-
|
|
126
|
-
const arweave = getArweave()
|
|
127
|
-
|
|
128
|
-
if (!arweave) {
|
|
129
|
-
logger('[fetchAll/actors] [fetchAllBinaryData] arweave not available')
|
|
130
|
-
return []
|
|
131
|
-
}
|
|
132
|
-
|
|
133
|
-
try {
|
|
134
|
-
const res = await fetch(
|
|
135
|
-
`https://${ARWEAVE_HOST}/tx/${transactionId}/status`,
|
|
136
|
-
)
|
|
137
|
-
|
|
138
|
-
if (res.status !== 200) {
|
|
139
|
-
logger(
|
|
140
|
-
`[fetchAll/actors] [fetchAllBinaryData] error fetching transaction data for ${transactionId}`,
|
|
141
|
-
)
|
|
142
|
-
|
|
143
|
-
excludedTransactions.add(transactionId)
|
|
144
|
-
|
|
145
|
-
await saveAppState(
|
|
146
|
-
'excludedTransactions',
|
|
147
|
-
JSON.stringify(Array.from(excludedTransactions)),
|
|
148
|
-
)
|
|
149
|
-
|
|
150
|
-
logger(
|
|
151
|
-
'[fetchAll/actors] [fetchAllBinaryData] updated excludedTransactions:',
|
|
152
|
-
excludedTransactions,
|
|
153
|
-
)
|
|
154
|
-
|
|
155
|
-
continue
|
|
156
|
-
}
|
|
157
|
-
|
|
158
|
-
const dataString = await arweave.transactions
|
|
159
|
-
.getData(transactionId, {
|
|
160
|
-
decode: true,
|
|
161
|
-
string: true,
|
|
162
|
-
})
|
|
163
|
-
.catch((error) => {
|
|
164
|
-
logger(
|
|
165
|
-
`[fetchAll/actors] [fetchAllBinaryData] error fetching transaction data for ${transactionId}`,
|
|
166
|
-
error,
|
|
167
|
-
)
|
|
168
|
-
})
|
|
169
|
-
|
|
170
|
-
const dataUint8Array = await arweave.transactions.getData(transactionId)
|
|
171
|
-
// let buffer
|
|
172
|
-
//
|
|
173
|
-
// if (dataUint8Array && dataUint8Array instanceof Uint8Array) {
|
|
174
|
-
// }
|
|
175
|
-
|
|
176
|
-
let contentType = identifyString(dataString)
|
|
177
|
-
if (
|
|
178
|
-
contentType !== 'json' &&
|
|
179
|
-
contentType !== 'base64' &&
|
|
180
|
-
contentType !== 'html'
|
|
181
|
-
) {
|
|
182
|
-
const possibleImageType = getDataTypeFromString(dataString)
|
|
183
|
-
if (!possibleImageType) {
|
|
184
|
-
logger(
|
|
185
|
-
`[fetchAll/actors] [fetchAllBinaryData] transaction ${transactionId} data not in expected format: ${possibleImageType}`,
|
|
186
|
-
)
|
|
187
|
-
continue
|
|
188
|
-
}
|
|
189
|
-
|
|
190
|
-
contentType = possibleImageType
|
|
191
|
-
}
|
|
192
|
-
|
|
193
|
-
if (contentType === 'url') {
|
|
194
|
-
const url = dataString as string
|
|
195
|
-
const response = await fetch(url)
|
|
196
|
-
if (!response.ok) {
|
|
197
|
-
throw new Error(`Failed to fetch image: ${response.statusText}`)
|
|
198
|
-
}
|
|
199
|
-
|
|
200
|
-
// Get the image as a Blob
|
|
201
|
-
const blob = await response.blob()
|
|
202
|
-
const buffer = await blob.arrayBuffer()
|
|
203
|
-
const bufferUint8Array = new Uint8Array(buffer)
|
|
204
|
-
|
|
205
|
-
// Extract the file extension from the URL
|
|
206
|
-
const extensionMatch = url.match(
|
|
207
|
-
/\.(jpg|jpeg|png|gif|bmp|webp|svg)$/i,
|
|
208
|
-
)
|
|
209
|
-
if (!extensionMatch) {
|
|
210
|
-
throw new Error(
|
|
211
|
-
'Unable to determine the file extension from the URL.',
|
|
212
|
-
)
|
|
213
|
-
}
|
|
214
|
-
const fileExtension = extensionMatch[0] // e.g., ".jpg"
|
|
215
|
-
|
|
216
|
-
// Set the file name (you can customize this)
|
|
217
|
-
// const fileNameFromUrl = `${transactionId}${fileExtension}`
|
|
218
|
-
|
|
219
|
-
await fs.promises.writeFile(
|
|
220
|
-
`/files/images/${transactionId}`,
|
|
221
|
-
bufferUint8Array,
|
|
222
|
-
{
|
|
223
|
-
encoding: 'binary',
|
|
224
|
-
},
|
|
225
|
-
)
|
|
226
|
-
|
|
227
|
-
continue
|
|
228
|
-
}
|
|
229
|
-
|
|
230
|
-
let mimeType = getMimeType(dataString as string)
|
|
231
|
-
let fileExtension = mimeType
|
|
232
|
-
|
|
233
|
-
if (fileExtension && fileExtension?.startsWith('image')) {
|
|
234
|
-
fileExtension = fileExtension.replace('image/', '')
|
|
235
|
-
}
|
|
236
|
-
|
|
237
|
-
let fileName = transactionId
|
|
238
|
-
|
|
239
|
-
if (contentType === 'base64') {
|
|
240
|
-
if (mimeType) {
|
|
241
|
-
fileName += `.${fileExtension}`
|
|
242
|
-
}
|
|
243
|
-
|
|
244
|
-
// Remove the Base64 header if it exists (e.g., "data:image/png;base64,")
|
|
245
|
-
const base64Data = dataString.split(',').pop() || ''
|
|
246
|
-
|
|
247
|
-
// Decode the Base64 string to binary
|
|
248
|
-
const binaryString = atob(base64Data)
|
|
249
|
-
const length = binaryString.length
|
|
250
|
-
const binaryData = new Uint8Array(length)
|
|
251
|
-
|
|
252
|
-
for (let i = 0; i < length; i++) {
|
|
253
|
-
binaryData[i] = binaryString.charCodeAt(i)
|
|
254
|
-
}
|
|
255
|
-
|
|
256
|
-
console.log(`attempting to writeFile to /files/images/${fileName}`)
|
|
257
|
-
|
|
258
|
-
await fs.promises.writeFile(`/files/images/${fileName}`, binaryData, {
|
|
259
|
-
encoding: 'binary',
|
|
260
|
-
})
|
|
261
|
-
|
|
262
|
-
// if (dataUint8Array && dataUint8Array instanceof Uint8Array) {
|
|
263
|
-
// await fs.promises.writeFile(
|
|
264
|
-
// `/files/images/${fileName}`,
|
|
265
|
-
// dataUint8Array,
|
|
266
|
-
// )
|
|
267
|
-
// }
|
|
268
|
-
}
|
|
29
|
+
fetchMetadata().then((metadataRecords) => {
|
|
30
|
+
sendBack({ type: 'fetchingMetadataSuccess', metadataRecords })
|
|
31
|
+
})
|
|
269
32
|
|
|
270
|
-
|
|
271
|
-
|
|
272
|
-
|
|
273
|
-
}
|
|
33
|
+
return () => {}
|
|
34
|
+
},
|
|
35
|
+
)
|
|
274
36
|
|
|
275
|
-
|
|
276
|
-
|
|
277
|
-
|
|
278
|
-
|
|
279
|
-
|
|
280
|
-
logger(error)
|
|
281
|
-
}
|
|
282
|
-
}
|
|
37
|
+
export const fetchBinaryData = fromCallback<
|
|
38
|
+
EventObject,
|
|
39
|
+
typeof downloadMachine
|
|
40
|
+
>(({ sendBack, receive, input: { context } }) => {
|
|
41
|
+
const { addresses } = context
|
|
283
42
|
|
|
43
|
+
const fetchBinaryData = async () => {
|
|
284
44
|
return []
|
|
285
45
|
}
|
|
286
46
|
|
|
287
|
-
|
|
288
|
-
sendBack({ type: '
|
|
47
|
+
fetchBinaryData().then(() => {
|
|
48
|
+
sendBack({ type: 'fetchingBinaryDataSuccess' })
|
|
289
49
|
})
|
|
290
50
|
|
|
291
51
|
return () => {}
|
|
@@ -70,13 +70,13 @@ const processUploadProperty = async (
|
|
|
70
70
|
for (const childProperty of uploadProperty.childProperties) {
|
|
71
71
|
const filePath = childProperty.localStoragePath
|
|
72
72
|
|
|
73
|
-
if (!filePath) {
|
|
73
|
+
if (!filePath || filePath.endsWith('undefined')) {
|
|
74
74
|
continue
|
|
75
75
|
}
|
|
76
76
|
|
|
77
77
|
const exists = await fs.promises.exists(filePath)
|
|
78
78
|
if (!exists) {
|
|
79
|
-
|
|
79
|
+
continue
|
|
80
80
|
}
|
|
81
81
|
|
|
82
82
|
childUploads.push({
|
|
@@ -113,6 +113,10 @@ const processUploadProperty = async (
|
|
|
113
113
|
if (relatedItemProperty && relatedItemProperty.localStoragePath) {
|
|
114
114
|
const filePath = relatedItemProperty.localStoragePath
|
|
115
115
|
|
|
116
|
+
if (!filePath || filePath.endsWith('undefined')) {
|
|
117
|
+
return uploads
|
|
118
|
+
}
|
|
119
|
+
|
|
116
120
|
const exists = await fs.promises.exists(filePath)
|
|
117
121
|
if (!exists) {
|
|
118
122
|
return uploads
|
package/dist/src/index.ts
CHANGED
|
@@ -1,73 +1,58 @@
|
|
|
1
1
|
import { assign, setup } from 'xstate'
|
|
2
2
|
import {
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
} from '@/browser/schema/file/
|
|
6
|
-
import { Attestation } from '@/browser/gql/graphql'
|
|
3
|
+
fetchBinaryData,
|
|
4
|
+
fetchMetadata,
|
|
5
|
+
} from '@/browser/schema/file/download/actors'
|
|
7
6
|
|
|
8
|
-
type
|
|
9
|
-
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
7
|
+
type DownloadMachineContext = {
|
|
8
|
+
fileName: string
|
|
9
|
+
metadata: any
|
|
10
|
+
binaryData: any
|
|
11
|
+
metadataServiceUrl: string
|
|
12
|
+
blobServiceUrl: string
|
|
13
13
|
}
|
|
14
14
|
|
|
15
|
-
export const
|
|
15
|
+
export const downloadMachine = setup({
|
|
16
16
|
types: {
|
|
17
|
-
context: {} as
|
|
17
|
+
context: {} as DownloadMachineContext,
|
|
18
18
|
},
|
|
19
19
|
actors: {
|
|
20
|
-
|
|
21
|
-
|
|
20
|
+
fetchMetadata,
|
|
21
|
+
fetchBinaryData,
|
|
22
22
|
},
|
|
23
23
|
}).createMachine({
|
|
24
|
-
id: '
|
|
25
|
-
initial: '
|
|
26
|
-
context: ({ input }) =>
|
|
27
|
-
({
|
|
28
|
-
...input,
|
|
29
|
-
dbsLoaded: false,
|
|
30
|
-
}) as FetchAllFilesMachineContext,
|
|
31
|
-
on: {
|
|
32
|
-
allDbsLoaded: {
|
|
33
|
-
target: '.fetchingAllMetadataRecords',
|
|
34
|
-
actions: assign({
|
|
35
|
-
dbsLoaded: true,
|
|
36
|
-
}),
|
|
37
|
-
},
|
|
38
|
-
},
|
|
24
|
+
id: 'download',
|
|
25
|
+
initial: 'fetchingMetadata',
|
|
26
|
+
context: ({ input }) => input as DownloadMachineContext,
|
|
39
27
|
states: {
|
|
40
28
|
idle: {},
|
|
41
|
-
|
|
29
|
+
fetchingMetadata: {
|
|
42
30
|
on: {
|
|
43
|
-
|
|
44
|
-
target: '
|
|
31
|
+
fetchingMetadataSuccess: {
|
|
32
|
+
target: 'fetchingBinaryData',
|
|
45
33
|
actions: assign({
|
|
46
|
-
|
|
34
|
+
metadata: (context, event) => event.metadataRecords,
|
|
47
35
|
}),
|
|
48
36
|
},
|
|
49
37
|
},
|
|
50
38
|
invoke: {
|
|
51
|
-
src: '
|
|
52
|
-
input: (
|
|
39
|
+
src: 'fetchMetadata',
|
|
40
|
+
input: (context) => ({ context }),
|
|
53
41
|
},
|
|
54
42
|
},
|
|
55
|
-
|
|
43
|
+
fetchingBinaryData: {
|
|
56
44
|
on: {
|
|
57
|
-
|
|
58
|
-
target: '
|
|
45
|
+
fetchingBinaryDataSuccess: {
|
|
46
|
+
target: 'idle',
|
|
59
47
|
actions: assign({
|
|
60
|
-
|
|
48
|
+
binaryData: (context, event) => event.data,
|
|
61
49
|
}),
|
|
62
50
|
},
|
|
63
51
|
},
|
|
64
52
|
invoke: {
|
|
65
|
-
src: '
|
|
66
|
-
input: (
|
|
53
|
+
src: 'fetchBinaryData',
|
|
54
|
+
input: (context) => ({ context }),
|
|
67
55
|
},
|
|
68
56
|
},
|
|
69
|
-
success: {
|
|
70
|
-
type: 'final',
|
|
71
|
-
},
|
|
72
57
|
},
|
|
73
58
|
})
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"getPublishUploads.d.ts","sourceRoot":"","sources":["../../../../../../src/browser/db/read/getPublishUploads.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,MAAM,WAAW,CAAA;AAC9C,OAAO,WAAW,MAAM,6BAA6B,CAAA;AAUrD,eAAO,MAAM,cAAc,WACjB,MAAM,GAAG,IAAI,GAAG,SAAS,QAC3B,UAAU,GAAG,SAAS,KAC3B,OAAO,CAAC,MAAM,CAkBhB,CAAA;AACD,eAAO,MAAM,yBAAyB,SAC9B,MAAM,GAAG,UAAU,eACZ,MAAM,GAAG,SAAS,KAC9B,OAAO,CAAC,WAAW,CAerB,CAAA;AACD,MAAM,MAAM,cAAc,GAAG;IAC3B,YAAY,EAAE,YAAY,CAAC,GAAG,CAAC,CAAA;IAC/B,eAAe,EAAE,YAAY,CAAC,GAAG,CAAC,EAAE,CAAA;CACrC,CAAA;
|
|
1
|
+
{"version":3,"file":"getPublishUploads.d.ts","sourceRoot":"","sources":["../../../../../../src/browser/db/read/getPublishUploads.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,IAAI,EAAE,YAAY,EAAE,MAAM,WAAW,CAAA;AAC9C,OAAO,WAAW,MAAM,6BAA6B,CAAA;AAUrD,eAAO,MAAM,cAAc,WACjB,MAAM,GAAG,IAAI,GAAG,SAAS,QAC3B,UAAU,GAAG,SAAS,KAC3B,OAAO,CAAC,MAAM,CAkBhB,CAAA;AACD,eAAO,MAAM,yBAAyB,SAC9B,MAAM,GAAG,UAAU,eACZ,MAAM,GAAG,SAAS,KAC9B,OAAO,CAAC,WAAW,CAerB,CAAA;AACD,MAAM,MAAM,cAAc,GAAG;IAC3B,YAAY,EAAE,YAAY,CAAC,GAAG,CAAC,CAAA;IAC/B,eAAe,EAAE,YAAY,CAAC,GAAG,CAAC,EAAE,CAAA;CACrC,CAAA;AAiID,MAAM,MAAM,aAAa,GAAG;IAC1B,gBAAgB,EAAE,MAAM,CAAA;IACxB,mBAAmB,EAAE,MAAM,CAAA;IAC3B,WAAW,EAAE,MAAM,CAAA;IACnB,cAAc,EAAE,MAAM,CAAA;IACtB,iBAAiB,EAAE,WAAW,CAAA;CAC/B,CAAA;AACD,eAAO,MAAM,iBAAiB,SACtB,IAAI,CAAC,GAAG,CAAC,YACN,aAAa,EAAE,wBACF,YAAY,CAAC,GAAG,CAAC,6BA+CxC,CAAA"}
|