@lumeweb/pinner 0.0.1 → 0.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (186) hide show
  1. package/LICENSE +9 -0
  2. package/README.md +690 -28
  3. package/dist/cjs/_virtual/rolldown_runtime.cjs +29 -0
  4. package/dist/cjs/adapters/pinata/adapter.cjs +88 -0
  5. package/dist/cjs/adapters/pinata/adapter.cjs.map +1 -0
  6. package/dist/cjs/adapters/pinata/adapter.d.cts +35 -0
  7. package/dist/cjs/adapters/pinata/builder.cjs +194 -0
  8. package/dist/cjs/adapters/pinata/builder.cjs.map +1 -0
  9. package/dist/cjs/adapters/pinata/index.cjs +3 -0
  10. package/dist/cjs/adapters/pinata/list-builder.cjs +52 -0
  11. package/dist/cjs/adapters/pinata/list-builder.cjs.map +1 -0
  12. package/dist/cjs/blockstore/index.cjs +2 -0
  13. package/dist/cjs/blockstore/unstorage-base.cjs +240 -0
  14. package/dist/cjs/blockstore/unstorage-base.cjs.map +1 -0
  15. package/dist/cjs/blockstore/unstorage-base.d.cts +23 -0
  16. package/dist/cjs/blockstore/unstorage.cjs +39 -0
  17. package/dist/cjs/blockstore/unstorage.cjs.map +1 -0
  18. package/dist/cjs/blockstore/unstorage.d.cts +36 -0
  19. package/dist/cjs/config.d.cts +51 -0
  20. package/dist/cjs/encoder/base64.cjs +38 -0
  21. package/dist/cjs/encoder/base64.cjs.map +1 -0
  22. package/dist/cjs/encoder/csv/csv-formatter.cjs +81 -0
  23. package/dist/cjs/encoder/csv/csv-formatter.cjs.map +1 -0
  24. package/dist/cjs/encoder/csv/field-formatter.cjs +76 -0
  25. package/dist/cjs/encoder/csv/field-formatter.cjs.map +1 -0
  26. package/dist/cjs/encoder/csv/row-formatter.cjs +159 -0
  27. package/dist/cjs/encoder/csv/row-formatter.cjs.map +1 -0
  28. package/dist/cjs/encoder/csv.cjs +44 -0
  29. package/dist/cjs/encoder/csv.cjs.map +1 -0
  30. package/dist/cjs/encoder/error.cjs +19 -0
  31. package/dist/cjs/encoder/error.cjs.map +1 -0
  32. package/dist/cjs/encoder/index.cjs +6 -0
  33. package/dist/cjs/encoder/json.cjs +36 -0
  34. package/dist/cjs/encoder/json.cjs.map +1 -0
  35. package/dist/cjs/encoder/text.cjs +35 -0
  36. package/dist/cjs/encoder/text.cjs.map +1 -0
  37. package/dist/cjs/encoder/url.cjs +39 -0
  38. package/dist/cjs/encoder/url.cjs.map +1 -0
  39. package/dist/cjs/errors/index.cjs +104 -0
  40. package/dist/cjs/errors/index.cjs.map +1 -0
  41. package/dist/cjs/errors/index.d.cts +47 -0
  42. package/dist/cjs/index.cjs +42 -0
  43. package/dist/cjs/index.d.cts +14 -0
  44. package/dist/cjs/pin/client.cjs +96 -0
  45. package/dist/cjs/pin/client.cjs.map +1 -0
  46. package/dist/cjs/pin/index.cjs +1 -0
  47. package/dist/cjs/pinner.cjs +126 -0
  48. package/dist/cjs/pinner.cjs.map +1 -0
  49. package/dist/cjs/pinner.d.cts +77 -0
  50. package/dist/cjs/types/constants.cjs +34 -0
  51. package/dist/cjs/types/constants.cjs.map +1 -0
  52. package/dist/cjs/types/mime-types.cjs +11 -0
  53. package/dist/cjs/types/mime-types.cjs.map +1 -0
  54. package/dist/cjs/types/mime-types.d.cts +7 -0
  55. package/dist/cjs/types/pin.d.cts +74 -0
  56. package/dist/cjs/types/pinata.d.cts +99 -0
  57. package/dist/cjs/types/type-guards.cjs +20 -0
  58. package/dist/cjs/types/type-guards.cjs.map +1 -0
  59. package/dist/cjs/types/type-guards.d.cts +15 -0
  60. package/dist/cjs/types/upload.cjs +18 -0
  61. package/dist/cjs/types/upload.cjs.map +1 -0
  62. package/dist/cjs/types/upload.d.cts +189 -0
  63. package/dist/cjs/upload/base-upload.cjs +135 -0
  64. package/dist/cjs/upload/base-upload.cjs.map +1 -0
  65. package/dist/cjs/upload/builder.cjs +174 -0
  66. package/dist/cjs/upload/builder.cjs.map +1 -0
  67. package/dist/cjs/upload/builder.d.cts +60 -0
  68. package/dist/cjs/upload/car.cjs +129 -0
  69. package/dist/cjs/upload/car.cjs.map +1 -0
  70. package/dist/cjs/upload/car.d.cts +19 -0
  71. package/dist/cjs/upload/constants.cjs +9 -0
  72. package/dist/cjs/upload/constants.cjs.map +1 -0
  73. package/dist/cjs/upload/index.cjs +8 -0
  74. package/dist/cjs/upload/manager.cjs +249 -0
  75. package/dist/cjs/upload/manager.cjs.map +1 -0
  76. package/dist/cjs/upload/manager.d.cts +35 -0
  77. package/dist/cjs/upload/normalize.cjs +28 -0
  78. package/dist/cjs/upload/normalize.cjs.map +1 -0
  79. package/dist/cjs/upload/tus-upload.cjs +74 -0
  80. package/dist/cjs/upload/tus-upload.cjs.map +1 -0
  81. package/dist/cjs/upload/xhr-upload.cjs +41 -0
  82. package/dist/cjs/upload/xhr-upload.cjs.map +1 -0
  83. package/dist/cjs/utils/env.cjs +12 -0
  84. package/dist/cjs/utils/env.cjs.map +1 -0
  85. package/dist/cjs/utils/stream.cjs +141 -0
  86. package/dist/cjs/utils/stream.cjs.map +1 -0
  87. package/dist/cjs/utils/stream.d.cts +23 -0
  88. package/dist/cjs/utils/tus-patch.cjs +50 -0
  89. package/dist/cjs/utils/tus-patch.cjs.map +1 -0
  90. package/dist/cjs/utils/validation.cjs +62 -0
  91. package/dist/cjs/utils/validation.cjs.map +1 -0
  92. package/dist/esm/_virtual/rolldown_runtime.js +8 -0
  93. package/dist/esm/adapters/pinata/adapter.d.ts +35 -0
  94. package/dist/esm/adapters/pinata/adapter.js +87 -0
  95. package/dist/esm/adapters/pinata/adapter.js.map +1 -0
  96. package/dist/esm/adapters/pinata/builder.d.ts +1 -0
  97. package/dist/esm/adapters/pinata/builder.js +187 -0
  98. package/dist/esm/adapters/pinata/builder.js.map +1 -0
  99. package/dist/esm/adapters/pinata/index.d.ts +4 -0
  100. package/dist/esm/adapters/pinata/index.js +3 -0
  101. package/dist/esm/adapters/pinata/list-builder.d.ts +1 -0
  102. package/dist/esm/adapters/pinata/list-builder.js +51 -0
  103. package/dist/esm/adapters/pinata/list-builder.js.map +1 -0
  104. package/dist/esm/blockstore/index.d.ts +2 -0
  105. package/dist/esm/blockstore/index.js +2 -0
  106. package/dist/esm/blockstore/unstorage-base.d.ts +23 -0
  107. package/dist/esm/blockstore/unstorage-base.js +231 -0
  108. package/dist/esm/blockstore/unstorage-base.js.map +1 -0
  109. package/dist/esm/blockstore/unstorage.d.ts +36 -0
  110. package/dist/esm/blockstore/unstorage.js +38 -0
  111. package/dist/esm/blockstore/unstorage.js.map +1 -0
  112. package/dist/esm/config.d.ts +51 -0
  113. package/dist/esm/encoder/base64.js +37 -0
  114. package/dist/esm/encoder/base64.js.map +1 -0
  115. package/dist/esm/encoder/csv/csv-formatter.js +81 -0
  116. package/dist/esm/encoder/csv/csv-formatter.js.map +1 -0
  117. package/dist/esm/encoder/csv/field-formatter.js +75 -0
  118. package/dist/esm/encoder/csv/field-formatter.js.map +1 -0
  119. package/dist/esm/encoder/csv/row-formatter.js +159 -0
  120. package/dist/esm/encoder/csv/row-formatter.js.map +1 -0
  121. package/dist/esm/encoder/csv.js +43 -0
  122. package/dist/esm/encoder/csv.js.map +1 -0
  123. package/dist/esm/encoder/error.js +18 -0
  124. package/dist/esm/encoder/error.js.map +1 -0
  125. package/dist/esm/encoder/index.js +6 -0
  126. package/dist/esm/encoder/json.js +35 -0
  127. package/dist/esm/encoder/json.js.map +1 -0
  128. package/dist/esm/encoder/text.js +34 -0
  129. package/dist/esm/encoder/text.js.map +1 -0
  130. package/dist/esm/encoder/url.js +36 -0
  131. package/dist/esm/encoder/url.js.map +1 -0
  132. package/dist/esm/errors/index.d.ts +47 -0
  133. package/dist/esm/errors/index.js +93 -0
  134. package/dist/esm/errors/index.js.map +1 -0
  135. package/dist/esm/index.d.ts +16 -0
  136. package/dist/esm/index.js +14 -0
  137. package/dist/esm/pin/client.js +95 -0
  138. package/dist/esm/pin/client.js.map +1 -0
  139. package/dist/esm/pin/index.js +1 -0
  140. package/dist/esm/pinner.d.ts +77 -0
  141. package/dist/esm/pinner.js +125 -0
  142. package/dist/esm/pinner.js.map +1 -0
  143. package/dist/esm/types/constants.js +29 -0
  144. package/dist/esm/types/constants.js.map +1 -0
  145. package/dist/esm/types/mime-types.d.ts +7 -0
  146. package/dist/esm/types/mime-types.js +8 -0
  147. package/dist/esm/types/mime-types.js.map +1 -0
  148. package/dist/esm/types/pin.d.ts +74 -0
  149. package/dist/esm/types/pinata.d.ts +99 -0
  150. package/dist/esm/types/type-guards.d.ts +15 -0
  151. package/dist/esm/types/type-guards.js +19 -0
  152. package/dist/esm/types/type-guards.js.map +1 -0
  153. package/dist/esm/types/upload.d.ts +189 -0
  154. package/dist/esm/types/upload.js +16 -0
  155. package/dist/esm/types/upload.js.map +1 -0
  156. package/dist/esm/upload/base-upload.js +132 -0
  157. package/dist/esm/upload/base-upload.js.map +1 -0
  158. package/dist/esm/upload/builder.d.ts +60 -0
  159. package/dist/esm/upload/builder.js +173 -0
  160. package/dist/esm/upload/builder.js.map +1 -0
  161. package/dist/esm/upload/car.d.ts +19 -0
  162. package/dist/esm/upload/car.js +125 -0
  163. package/dist/esm/upload/car.js.map +1 -0
  164. package/dist/esm/upload/constants.js +7 -0
  165. package/dist/esm/upload/constants.js.map +1 -0
  166. package/dist/esm/upload/index.js +8 -0
  167. package/dist/esm/upload/manager.d.ts +35 -0
  168. package/dist/esm/upload/manager.js +248 -0
  169. package/dist/esm/upload/manager.js.map +1 -0
  170. package/dist/esm/upload/normalize.js +28 -0
  171. package/dist/esm/upload/normalize.js.map +1 -0
  172. package/dist/esm/upload/tus-upload.js +72 -0
  173. package/dist/esm/upload/tus-upload.js.map +1 -0
  174. package/dist/esm/upload/xhr-upload.js +39 -0
  175. package/dist/esm/upload/xhr-upload.js.map +1 -0
  176. package/dist/esm/utils/env.js +11 -0
  177. package/dist/esm/utils/env.js.map +1 -0
  178. package/dist/esm/utils/stream.d.ts +23 -0
  179. package/dist/esm/utils/stream.js +134 -0
  180. package/dist/esm/utils/stream.js.map +1 -0
  181. package/dist/esm/utils/tus-patch.js +51 -0
  182. package/dist/esm/utils/tus-patch.js.map +1 -0
  183. package/dist/esm/utils/validation.js +60 -0
  184. package/dist/esm/utils/validation.js.map +1 -0
  185. package/package.json +95 -8
  186. package/public/mockServiceWorker.js +349 -0
@@ -0,0 +1,41 @@
1
+ const require_rolldown_runtime = require('../_virtual/rolldown_runtime.cjs');
2
+ const require_upload = require('../types/upload.cjs');
3
+ const require_constants = require('./constants.cjs');
4
+ const require_base_upload = require('./base-upload.cjs');
5
+ let _lumeweb_uppy_post_upload = require("@lumeweb/uppy-post-upload");
6
+ _lumeweb_uppy_post_upload = require_rolldown_runtime.__toESM(_lumeweb_uppy_post_upload);
7
+
8
+ //#region src/upload/xhr-upload.ts
9
+ var XHRUploadHandler = class extends require_base_upload.BaseUploadHandler {
10
+ configurePlugin(uppy) {
11
+ uppy.use(_lumeweb_uppy_post_upload.default, {
12
+ endpoint: `${this.config.endpoint}/api/upload`,
13
+ fieldName: "file",
14
+ formData: true,
15
+ headers: { Authorization: `Bearer ${this.config.jwt}` }
16
+ });
17
+ }
18
+ parseResult(result) {
19
+ const uppyResponse = result;
20
+ const response = uppyResponse.body || uppyResponse;
21
+ return {
22
+ id: response.id,
23
+ cid: response.cid,
24
+ name: response.name,
25
+ size: response.size,
26
+ mimeType: response.mimeType,
27
+ createdAt: new Date(response.createdAt),
28
+ numberOfFiles: response.numberOfFiles,
29
+ keyvalues: response.keyvalues,
30
+ operationId: response.operationId,
31
+ [require_upload.UploadResultSymbol]: true
32
+ };
33
+ }
34
+ getUploadSource() {
35
+ return require_constants.UPLOAD_SOURCE_XHR;
36
+ }
37
+ };
38
+
39
+ //#endregion
40
+ exports.XHRUploadHandler = XHRUploadHandler;
41
+ //# sourceMappingURL=xhr-upload.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"xhr-upload.cjs","names":["BaseUploadHandler","XHRUpload","UploadResultSymbol","UPLOAD_SOURCE_XHR"],"sources":["../../../src/upload/xhr-upload.ts"],"sourcesContent":["import Uppy from \"@uppy/core\";\nimport XHRUpload from \"@lumeweb/uppy-post-upload\";\nimport type { UploadResult } from \"@/types/upload\";\nimport { UploadResultSymbol } from \"@/types/upload\";\nimport { BaseUploadHandler } from \"./base-upload\";\nimport { UPLOAD_SOURCE_XHR } from \"./constants\";\n\nexport class XHRUploadHandler extends BaseUploadHandler {\n protected configurePlugin(uppy: Uppy): void {\n uppy.use(XHRUpload, {\n endpoint: `${this.config.endpoint}/api/upload`,\n fieldName: \"file\",\n formData: true,\n headers: {\n Authorization: `Bearer ${this.config.jwt}`,\n },\n });\n }\n\n protected parseResult(result: unknown): UploadResult {\n const uppyResponse = result as {\n uploadURL: string;\n body?: {\n id: string;\n cid: string;\n name: string;\n size: number;\n mimeType: string;\n createdAt: string;\n numberOfFiles: number;\n keyvalues?: Record<string, string>;\n operationId?: number;\n };\n };\n\n const response = uppyResponse.body || (uppyResponse as any);\n\n return {\n id: response.id,\n cid: response.cid,\n name: response.name,\n size: response.size,\n mimeType: response.mimeType,\n createdAt: new Date(response.createdAt),\n numberOfFiles: response.numberOfFiles,\n keyvalues: response.keyvalues,\n operationId: response.operationId,\n [UploadResultSymbol]: true,\n };\n }\n\n protected getUploadSource(): string {\n return UPLOAD_SOURCE_XHR;\n }\n}\n"],"mappings":";;;;;;;;AAOA,IAAa,mBAAb,cAAsCA,sCAAkB;CACtD,AAAU,gBAAgB,MAAkB;AAC1C,OAAK,IAAIC,mCAAW;GAClB,UAAU,GAAG,KAAK,OAAO,SAAS;GAClC,WAAW;GACX,UAAU;GACV,SAAS,EACP,eAAe,UAAU,KAAK,OAAO,OACtC;GACF,CAAC;;CAGJ,AAAU,YAAY,QAA+B;EACnD,MAAM,eAAe;EAerB,MAAM,WAAW,aAAa,QAAS;AAEvC,SAAO;GACL,IAAI,SAAS;GACb,KAAK,SAAS;GACd,MAAM,SAAS;GACf,MAAM,SAAS;GACf,UAAU,SAAS;GACnB,WAAW,IAAI,KAAK,SAAS,UAAU;GACvC,eAAe,SAAS;GACxB,WAAW,SAAS;GACpB,aAAa,SAAS;IACrBC,oCAAqB;GACvB;;CAGH,AAAU,kBAA0B;AAClC,SAAOC"}
@@ -0,0 +1,12 @@
1
+
2
+ //#region src/utils/env.ts
3
+ /**
4
+ * Check if the current environment is Node.js.
5
+ */
6
+ function isNodeEnvironment() {
7
+ return typeof process !== "undefined" && process?.versions?.node !== void 0;
8
+ }
9
+
10
+ //#endregion
11
+ exports.isNodeEnvironment = isNodeEnvironment;
12
+ //# sourceMappingURL=env.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"env.cjs","names":[],"sources":["../../../src/utils/env.ts"],"sourcesContent":["/**\n * Check if the current environment is Node.js.\n */\nexport function isNodeEnvironment(): boolean {\n return typeof process !== 'undefined' && process?.versions?.node !== undefined;\n}\n"],"mappings":";;;;;AAGA,SAAgB,oBAA6B;AAC3C,QAAO,OAAO,YAAY,eAAe,SAAS,UAAU,SAAS"}
@@ -0,0 +1,141 @@
1
+ const require_env = require('./env.cjs');
2
+
3
+ //#region src/utils/stream.ts
4
+ /**
5
+ * Convert a ReadableStream to a Blob.
6
+ */
7
+ async function streamToBlob(stream, mimeType) {
8
+ const chunks = [];
9
+ const reader = stream.getReader();
10
+ while (true) {
11
+ const { done, value } = await reader.read();
12
+ if (done) break;
13
+ chunks.push(value);
14
+ }
15
+ return new Blob(chunks, { type: mimeType });
16
+ }
17
+ /**
18
+ * Calculate the total size of a ReadableStream by consuming it.
19
+ */
20
+ async function calculateStreamSize(stream, signal) {
21
+ let size = 0n;
22
+ const reader = stream.getReader();
23
+ while (true) {
24
+ if (signal?.aborted) throw new Error("Aborted");
25
+ const { done, value } = await reader.read();
26
+ if (done) break;
27
+ size += BigInt(value.length);
28
+ }
29
+ return size;
30
+ }
31
+ /**
32
+ * Convert an async generator to a ReadableStream.
33
+ */
34
+ function asyncGeneratorToReadableStream(generator) {
35
+ return new ReadableStream({ async start(controller) {
36
+ try {
37
+ for await (const item of generator) controller.enqueue(item);
38
+ controller.close();
39
+ } catch (error) {
40
+ controller.error(error);
41
+ }
42
+ } });
43
+ }
44
+ /**
45
+ * Convert a ReadableStream to an async iterable.
46
+ */
47
+ async function* readableStreamToAsyncIterable(stream) {
48
+ const reader = stream.getReader();
49
+ try {
50
+ while (true) {
51
+ const { done, value } = await reader.read();
52
+ if (done) break;
53
+ yield value;
54
+ }
55
+ } finally {
56
+ reader.releaseLock();
57
+ }
58
+ }
59
+ /**
60
+ * Convert a web ReadableStream to a Node.js stream.Readable
61
+ * This is needed for Node.js environments where tus-js-client expects Node streams.
62
+ */
63
+ async function readableStreamToNodeStream(stream) {
64
+ if (!require_env.isNodeEnvironment()) throw new Error("readableStreamToNodeStream can only be used in Node.js environment");
65
+ const { Readable } = await import("stream");
66
+ const reader = stream.getReader();
67
+ return new Readable({
68
+ async read() {
69
+ try {
70
+ const { done, value } = await reader.read();
71
+ if (done) {
72
+ reader.releaseLock();
73
+ this.push(null);
74
+ } else this.push(Buffer.from(value));
75
+ } catch (error) {
76
+ reader.releaseLock();
77
+ this.destroy(error);
78
+ }
79
+ },
80
+ destroy(error, callback) {
81
+ try {
82
+ reader.releaseLock();
83
+ } catch (e) {}
84
+ if (callback) callback(error);
85
+ }
86
+ });
87
+ }
88
+ /**
89
+ * Convert a ReadableStream to a Blob using the Response API.
90
+ * This is the preferred method in browser environments as it's built-in and efficient.
91
+ */
92
+ async function streamToBlobViaResponse(stream) {
93
+ return new Response(stream).blob();
94
+ }
95
+ /**
96
+ * Convert a File to a ReadableStream of Uint8Array without loading entire blob into memory.
97
+ * This streams the file content chunk by chunk.
98
+ */
99
+ function fileToReadableStream(file) {
100
+ return new ReadableStream({ async start(controller) {
101
+ try {
102
+ const reader = file.stream().getReader();
103
+ while (true) {
104
+ const { done, value } = await reader.read();
105
+ if (done) {
106
+ controller.close();
107
+ break;
108
+ }
109
+ controller.enqueue(value);
110
+ }
111
+ } catch (error) {
112
+ controller.error(error);
113
+ }
114
+ } });
115
+ }
116
+ /**
117
+ * Collect all chunks from an async iterable or iterable into a single Uint8Array.
118
+ */
119
+ async function collectAsyncIterable(iterable) {
120
+ const chunks = [];
121
+ for await (const chunk of iterable) chunks.push(chunk);
122
+ const totalLength = chunks.reduce((acc, chunk) => acc + chunk.length, 0);
123
+ const result = new Uint8Array(totalLength);
124
+ let offset = 0;
125
+ for (const chunk of chunks) {
126
+ result.set(chunk, offset);
127
+ offset += chunk.length;
128
+ }
129
+ return result;
130
+ }
131
+
132
+ //#endregion
133
+ exports.asyncGeneratorToReadableStream = asyncGeneratorToReadableStream;
134
+ exports.calculateStreamSize = calculateStreamSize;
135
+ exports.collectAsyncIterable = collectAsyncIterable;
136
+ exports.fileToReadableStream = fileToReadableStream;
137
+ exports.readableStreamToAsyncIterable = readableStreamToAsyncIterable;
138
+ exports.readableStreamToNodeStream = readableStreamToNodeStream;
139
+ exports.streamToBlob = streamToBlob;
140
+ exports.streamToBlobViaResponse = streamToBlobViaResponse;
141
+ //# sourceMappingURL=stream.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"stream.cjs","names":["isNodeEnvironment"],"sources":["../../../src/utils/stream.ts"],"sourcesContent":["import type { AwaitIterable } from \"interface-store\";\nimport { isNodeEnvironment } from \"./env\";\n\n/**\n * Convert a ReadableStream to a Blob.\n */\nexport async function streamToBlob(\n stream: ReadableStream<Uint8Array>,\n mimeType: string,\n): Promise<Blob> {\n const chunks: Uint8Array[] = [];\n const reader = stream.getReader();\n\n while (true) {\n const { done, value } = await reader.read();\n if (done) break;\n chunks.push(value);\n }\n\n return new Blob(chunks as any, { type: mimeType });\n}\n\n/**\n * Calculate the total size of a ReadableStream by consuming it.\n */\nexport async function calculateStreamSize(\n stream: ReadableStream<Uint8Array>,\n signal?: AbortSignal,\n): Promise<bigint> {\n let size = 0n;\n const reader = stream.getReader();\n\n while (true) {\n if (signal?.aborted) {\n throw new Error(\"Aborted\");\n }\n\n const { done, value } = await reader.read();\n if (done) break;\n size += BigInt(value.length);\n }\n\n return size;\n}\n\n/**\n * Convert an async generator to a ReadableStream.\n */\nexport function asyncGeneratorToReadableStream<T>(\n generator: AsyncGenerator<T>,\n): ReadableStream<T> {\n return new ReadableStream({\n async start(controller) {\n try {\n for await (const item of generator) {\n controller.enqueue(item);\n }\n controller.close();\n } catch (error) {\n controller.error(error);\n }\n },\n });\n}\n\n/**\n * Convert a ReadableStream to an async iterable.\n */\nexport async function* readableStreamToAsyncIterable<T>(\n stream: ReadableStream<T>,\n): AsyncIterable<T> {\n const reader = stream.getReader();\n\n try {\n while (true) {\n const { done, value } = await reader.read();\n if (done) break;\n yield value;\n }\n } finally {\n reader.releaseLock();\n }\n}\n\n/**\n * Convert a web ReadableStream to a Node.js stream.Readable\n * This is needed for Node.js environments where tus-js-client expects Node streams.\n */\nexport async function readableStreamToNodeStream(\n stream: ReadableStream<Uint8Array>,\n): Promise<import(\"stream\").Readable> {\n if (!isNodeEnvironment()) {\n throw new Error(\n \"readableStreamToNodeStream can only be used in Node.js environment\",\n );\n }\n\n const { Readable } = await import(\"stream\");\n\n // Create a single reader for the entire stream lifecycle\n // The read() method is called multiple times by Node.js, so we cannot\n // call getReader() inside it - that would create multiple readers and\n // cause \"ReadableStream is locked\" errors\n const reader = stream.getReader();\n\n return new Readable({\n async read() {\n try {\n const { done, value } = await reader.read();\n if (done) {\n reader.releaseLock();\n this.push(null);\n } else {\n this.push(Buffer.from(value));\n }\n } catch (error) {\n reader.releaseLock();\n this.destroy(error as Error);\n }\n },\n\n // Ensure reader is released if stream is destroyed\n destroy(error, callback) {\n try {\n reader.releaseLock();\n } catch (e) {\n // Ignore errors during cleanup\n }\n if (callback) callback(error);\n },\n });\n}\n\n/**\n * Convert a ReadableStream to a Blob using the Response API.\n * This is the preferred method in browser environments as it's built-in and efficient.\n */\nexport async function streamToBlobViaResponse(\n stream: ReadableStream<Uint8Array>,\n): Promise<Blob> {\n return new Response(stream).blob();\n}\n\n/**\n * Convert a File to a ReadableStream of Uint8Array without loading entire blob into memory.\n * This streams the file content chunk by chunk.\n */\nexport function fileToReadableStream(file: File): ReadableStream<Uint8Array> {\n return new ReadableStream<Uint8Array>({\n async start(controller) {\n try {\n const reader = (file as any).stream().getReader();\n\n while (true) {\n const { done, value } = await reader.read();\n if (done) {\n controller.close();\n break;\n }\n controller.enqueue(value);\n }\n } catch (error) {\n controller.error(error);\n }\n },\n });\n}\n\n/**\n * Collect all chunks from an async iterable or iterable into a single Uint8Array.\n */\nexport async function collectAsyncIterable(\n iterable: AwaitIterable<Uint8Array>,\n): Promise<Uint8Array> {\n const chunks: Uint8Array[] = [];\n for await (const chunk of iterable) {\n chunks.push(chunk);\n }\n\n const totalLength = chunks.reduce((acc, chunk) => acc + chunk.length, 0);\n const result = new Uint8Array(totalLength);\n let offset = 0;\n for (const chunk of chunks) {\n result.set(chunk, offset);\n offset += chunk.length;\n }\n\n return result;\n}\n"],"mappings":";;;;;;AAMA,eAAsB,aACpB,QACA,UACe;CACf,MAAM,SAAuB,EAAE;CAC/B,MAAM,SAAS,OAAO,WAAW;AAEjC,QAAO,MAAM;EACX,MAAM,EAAE,MAAM,UAAU,MAAM,OAAO,MAAM;AAC3C,MAAI,KAAM;AACV,SAAO,KAAK,MAAM;;AAGpB,QAAO,IAAI,KAAK,QAAe,EAAE,MAAM,UAAU,CAAC;;;;;AAMpD,eAAsB,oBACpB,QACA,QACiB;CACjB,IAAI,OAAO;CACX,MAAM,SAAS,OAAO,WAAW;AAEjC,QAAO,MAAM;AACX,MAAI,QAAQ,QACV,OAAM,IAAI,MAAM,UAAU;EAG5B,MAAM,EAAE,MAAM,UAAU,MAAM,OAAO,MAAM;AAC3C,MAAI,KAAM;AACV,UAAQ,OAAO,MAAM,OAAO;;AAG9B,QAAO;;;;;AAMT,SAAgB,+BACd,WACmB;AACnB,QAAO,IAAI,eAAe,EACxB,MAAM,MAAM,YAAY;AACtB,MAAI;AACF,cAAW,MAAM,QAAQ,UACvB,YAAW,QAAQ,KAAK;AAE1B,cAAW,OAAO;WACX,OAAO;AACd,cAAW,MAAM,MAAM;;IAG5B,CAAC;;;;;AAMJ,gBAAuB,8BACrB,QACkB;CAClB,MAAM,SAAS,OAAO,WAAW;AAEjC,KAAI;AACF,SAAO,MAAM;GACX,MAAM,EAAE,MAAM,UAAU,MAAM,OAAO,MAAM;AAC3C,OAAI,KAAM;AACV,SAAM;;WAEA;AACR,SAAO,aAAa;;;;;;;AAQxB,eAAsB,2BACpB,QACoC;AACpC,KAAI,CAACA,+BAAmB,CACtB,OAAM,IAAI,MACR,qEACD;CAGH,MAAM,EAAE,aAAa,MAAM,OAAO;CAMlC,MAAM,SAAS,OAAO,WAAW;AAEjC,QAAO,IAAI,SAAS;EAClB,MAAM,OAAO;AACX,OAAI;IACF,MAAM,EAAE,MAAM,UAAU,MAAM,OAAO,MAAM;AAC3C,QAAI,MAAM;AACR,YAAO,aAAa;AACpB,UAAK,KAAK,KAAK;UAEf,MAAK,KAAK,OAAO,KAAK,MAAM,CAAC;YAExB,OAAO;AACd,WAAO,aAAa;AACpB,SAAK,QAAQ,MAAe;;;EAKhC,QAAQ,OAAO,UAAU;AACvB,OAAI;AACF,WAAO,aAAa;YACb,GAAG;AAGZ,OAAI,SAAU,UAAS,MAAM;;EAEhC,CAAC;;;;;;AAOJ,eAAsB,wBACpB,QACe;AACf,QAAO,IAAI,SAAS,OAAO,CAAC,MAAM;;;;;;AAOpC,SAAgB,qBAAqB,MAAwC;AAC3E,QAAO,IAAI,eAA2B,EACpC,MAAM,MAAM,YAAY;AACtB,MAAI;GACF,MAAM,SAAU,KAAa,QAAQ,CAAC,WAAW;AAEjD,UAAO,MAAM;IACX,MAAM,EAAE,MAAM,UAAU,MAAM,OAAO,MAAM;AAC3C,QAAI,MAAM;AACR,gBAAW,OAAO;AAClB;;AAEF,eAAW,QAAQ,MAAM;;WAEpB,OAAO;AACd,cAAW,MAAM,MAAM;;IAG5B,CAAC;;;;;AAMJ,eAAsB,qBACpB,UACqB;CACrB,MAAM,SAAuB,EAAE;AAC/B,YAAW,MAAM,SAAS,SACxB,QAAO,KAAK,MAAM;CAGpB,MAAM,cAAc,OAAO,QAAQ,KAAK,UAAU,MAAM,MAAM,QAAQ,EAAE;CACxE,MAAM,SAAS,IAAI,WAAW,YAAY;CAC1C,IAAI,SAAS;AACb,MAAK,MAAM,SAAS,QAAQ;AAC1B,SAAO,IAAI,OAAO,OAAO;AACzB,YAAU,MAAM;;AAGlB,QAAO"}
@@ -0,0 +1,23 @@
1
+ import "stream";
2
+
3
+ //#region src/utils/stream.d.ts
4
+
5
+ /**
6
+ * Convert a ReadableStream to a Blob.
7
+ */
8
+ declare function streamToBlob(stream: ReadableStream<Uint8Array>, mimeType: string): Promise<Blob>;
9
+ /**
10
+ * Calculate the total size of a ReadableStream by consuming it.
11
+ */
12
+ declare function calculateStreamSize(stream: ReadableStream<Uint8Array>, signal?: AbortSignal): Promise<bigint>;
13
+ /**
14
+ * Convert an async generator to a ReadableStream.
15
+ */
16
+ declare function asyncGeneratorToReadableStream<T>(generator: AsyncGenerator<T>): ReadableStream<T>;
17
+ /**
18
+ * Convert a ReadableStream to an async iterable.
19
+ */
20
+ declare function readableStreamToAsyncIterable<T>(stream: ReadableStream<T>): AsyncIterable<T>;
21
+ //#endregion
22
+ export { asyncGeneratorToReadableStream, calculateStreamSize, readableStreamToAsyncIterable, streamToBlob };
23
+ //# sourceMappingURL=stream.d.cts.map
@@ -0,0 +1,50 @@
1
+ const require_env = require('./env.cjs');
2
+
3
+ //#region src/utils/tus-patch.ts
4
+ /**
5
+ * Runtime patch for tus-js-client's NodeHttpStack Request class.
6
+ *
7
+ * PROBLEM:
8
+ * Code using onBeforeRequest hook may need to call abort() on the request
9
+ * before send() is called. However, the Request class only sets this._request
10
+ * inside the send() method, so getUnderlyingObject() returns null before send().
11
+ *
12
+ * SOLUTION:
13
+ * Patch the Request class to initialize a dummy _request object with an abort()
14
+ * method immediately when the request is created.
15
+ *
16
+ * NOTE:
17
+ * This patch is only needed in Node.js environments. In browser environments,
18
+ * the tus-js-client uses a different HTTP stack that doesn't have this issue.
19
+ */
20
+ function patchTusNodeHttpStack() {
21
+ if (!require_env.isNodeEnvironment()) return;
22
+ try {
23
+ const tusHttpStackModule = require("tus-js-client/lib.es5/node/httpStack");
24
+ if (!tusHttpStackModule || !tusHttpStackModule.default) {
25
+ console.warn("[tus-patch] tus-js-client NodeHttpStack not found, patch skipped");
26
+ return;
27
+ }
28
+ const NodeHttpStack = tusHttpStackModule.default;
29
+ const originalCreateRequest = NodeHttpStack.prototype.createRequest;
30
+ if (typeof originalCreateRequest !== "function") {
31
+ console.warn("[tus-patch] createRequest method not found, patch skipped");
32
+ return;
33
+ }
34
+ NodeHttpStack.prototype.createRequest = function(method, url) {
35
+ const request = originalCreateRequest.call(this, method, url);
36
+ request._request = {
37
+ abort: () => {},
38
+ destroyed: false
39
+ };
40
+ return request;
41
+ };
42
+ console.debug("[tus-patch] Successfully patched tus-js-client NodeHttpStack");
43
+ } catch (error) {
44
+ console.warn("[tus-patch] Failed to patch tus-js-client:", error);
45
+ }
46
+ }
47
+
48
+ //#endregion
49
+ exports.patchTusNodeHttpStack = patchTusNodeHttpStack;
50
+ //# sourceMappingURL=tus-patch.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"tus-patch.cjs","names":["isNodeEnvironment"],"sources":["../../../src/utils/tus-patch.ts"],"sourcesContent":["/**\n * Runtime patch for tus-js-client's NodeHttpStack Request class.\n *\n * PROBLEM:\n * Code using onBeforeRequest hook may need to call abort() on the request\n * before send() is called. However, the Request class only sets this._request\n * inside the send() method, so getUnderlyingObject() returns null before send().\n *\n * SOLUTION:\n * Patch the Request class to initialize a dummy _request object with an abort()\n * method immediately when the request is created.\n *\n * NOTE:\n * This patch is only needed in Node.js environments. In browser environments,\n * the tus-js-client uses a different HTTP stack that doesn't have this issue.\n */\n\nimport { isNodeEnvironment } from \"./env\";\n\n// Track whether the patch has been applied to prevent multiple patches\nlet isPatched = false;\n\nexport function patchTusNodeHttpStack(): void {\n // Prevent multiple patches\n if (isPatched) {\n return;\n }\n\n // Only apply patch in Node.js environments\n if (!isNodeEnvironment()) {\n return;\n }\n\n // Find the tus-js-client NodeHttpStack module\n // It exports the default as NodeHttpStack with a nested Request class\n try {\n // Use dynamic require() to avoid bundling issues\n // eslint-disable-next-line @typescript-eslint/no-var-requires\n const tusHttpStackModule = require(\"tus-js-client/lib.es5/node/httpStack\");\n\n if (!tusHttpStackModule || !tusHttpStackModule.default) {\n console.warn(\n \"[tus-patch] tus-js-client NodeHttpStack not found, patch skipped\",\n );\n return;\n }\n\n const NodeHttpStack = tusHttpStackModule.default;\n\n // Get the original createRequest method\n const originalCreateRequest = NodeHttpStack.prototype.createRequest;\n\n if (typeof originalCreateRequest !== \"function\") {\n console.warn(\"[tus-patch] createRequest method not found, patch skipped\");\n return;\n }\n\n // Patch createRequest to set up a dummy _request on the Request instance\n NodeHttpStack.prototype.createRequest = function (\n method: string,\n url: string,\n ) {\n // Call the original createRequest to get the Request instance\n const request = originalCreateRequest.call(this, method, url);\n\n // Set a dummy _request object with an abort() method\n // This allows getUnderlyingObject().abort() to work before send() is called\n request._request = {\n abort: () => {\n // No-op abort before actual request is created\n // The real request will be created in send() and this will be replaced\n },\n // Preserve any existing properties that might be checked\n destroyed: false,\n };\n\n return request;\n };\n\n console.debug(\n \"[tus-patch] Successfully patched tus-js-client NodeHttpStack\",\n );\n } catch (error) {\n console.warn(\"[tus-patch] Failed to patch tus-js-client:\", error);\n }\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;AAsBA,SAAgB,wBAA8B;AAO5C,KAAI,CAACA,+BAAmB,CACtB;AAKF,KAAI;EAGF,MAAM,qBAAqB,QAAQ,uCAAuC;AAE1E,MAAI,CAAC,sBAAsB,CAAC,mBAAmB,SAAS;AACtD,WAAQ,KACN,mEACD;AACD;;EAGF,MAAM,gBAAgB,mBAAmB;EAGzC,MAAM,wBAAwB,cAAc,UAAU;AAEtD,MAAI,OAAO,0BAA0B,YAAY;AAC/C,WAAQ,KAAK,4DAA4D;AACzE;;AAIF,gBAAc,UAAU,gBAAgB,SACtC,QACA,KACA;GAEA,MAAM,UAAU,sBAAsB,KAAK,MAAM,QAAQ,IAAI;AAI7D,WAAQ,WAAW;IACjB,aAAa;IAKb,WAAW;IACZ;AAED,UAAO;;AAGT,UAAQ,MACN,+DACD;UACM,OAAO;AACd,UAAQ,KAAK,8CAA8C,MAAM"}
@@ -0,0 +1,62 @@
1
+ const require_rolldown_runtime = require('../_virtual/rolldown_runtime.cjs');
2
+ const require_index = require('../errors/index.cjs');
3
+ let ipaddr_js = require("ipaddr.js");
4
+ ipaddr_js = require_rolldown_runtime.__toESM(ipaddr_js);
5
+
6
+ //#region src/utils/validation.ts
7
+ /**
8
+ * URL validation utilities to prevent SSRF and other security issues.
9
+ */
10
+ const BLOCKED_IPV4_RANGES = new Set([
11
+ "private",
12
+ "loopback",
13
+ "linkLocal",
14
+ "reserved",
15
+ "broadcast",
16
+ "carrierGradeNat",
17
+ "unspecified"
18
+ ]);
19
+ const BLOCKED_IPV6_RANGES = new Set([
20
+ "uniqueLocal",
21
+ "loopback",
22
+ "linkLocal",
23
+ "reserved",
24
+ "multicast",
25
+ "ipv4Mapped",
26
+ "unspecified"
27
+ ]);
28
+ /**
29
+ * Validates a URL string to ensure it's safe to fetch.
30
+ * Only allows HTTP/HTTPS protocols and validates the URL format.
31
+ *
32
+ * @param urlString - The URL string to validate
33
+ * @throws ValidationError if the URL is invalid or uses an unsafe protocol
34
+ */
35
+ function validateUrl(urlString) {
36
+ if (/\s/.test(urlString)) throw new require_index.ValidationError(`Invalid URL: contains whitespace characters`, "url");
37
+ try {
38
+ const url = new URL(urlString);
39
+ if (url.protocol !== "http:" && url.protocol !== "https:") throw new require_index.ValidationError(`Invalid URL protocol: ${url.protocol}. Only http: and https: are allowed.`, "url");
40
+ const hostname = url.hostname.toLowerCase();
41
+ if (hostname === "localhost") throw new require_index.ValidationError("Access to localhost addresses is not allowed", "url");
42
+ const cleanHostname = hostname.replace(/^\[|\]$/g, "");
43
+ if (ipaddr_js.default.isValid(cleanHostname)) {
44
+ const addr = ipaddr_js.default.parse(cleanHostname);
45
+ if (addr.kind() === "ipv4") {
46
+ const range = addr.range();
47
+ if (BLOCKED_IPV4_RANGES.has(range)) throw new require_index.ValidationError("Access to private IP addresses is not allowed", "url");
48
+ }
49
+ if (addr.kind() === "ipv6") {
50
+ const range = addr.range();
51
+ if (BLOCKED_IPV6_RANGES.has(range)) throw new require_index.ValidationError("Access to private IP addresses is not allowed", "url");
52
+ }
53
+ }
54
+ } catch (error) {
55
+ if (error instanceof require_index.ValidationError) throw error;
56
+ throw new require_index.ValidationError(`Invalid URL format: ${urlString}`, "url", error);
57
+ }
58
+ }
59
+
60
+ //#endregion
61
+ exports.validateUrl = validateUrl;
62
+ //# sourceMappingURL=validation.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"validation.cjs","names":["ValidationError","ipaddr"],"sources":["../../../src/utils/validation.ts"],"sourcesContent":["/**\n * URL validation utilities to prevent SSRF and other security issues.\n */\n\nimport { ValidationError } from \"@/errors\";\nimport ipaddr from \"ipaddr.js\";\n\nexport { ValidationError };\n\n// IPv4 ranges to block\nconst BLOCKED_IPV4_RANGES = new Set([\n \"private\", // 10.0.0.0/8, 172.16.0.0/12, 192.168.0.0/16\n \"loopback\", // 127.0.0.0/8\n \"linkLocal\", // 169.254.0.0/16\n \"reserved\",\n \"broadcast\",\n \"carrierGradeNat\", // 100.64.0.0/10\n \"unspecified\", // 0.0.0.0/8\n]);\n\n// IPv6 ranges to block\nconst BLOCKED_IPV6_RANGES = new Set([\n \"uniqueLocal\", // fc00::/7\n \"loopback\", // ::1\n \"linkLocal\", // fe80::/10\n \"reserved\",\n \"multicast\", // ff00::/8\n \"ipv4Mapped\", // ::ffff:0:0/96\n \"unspecified\", // ::\n]);\n\n/**\n * Validates a URL string to ensure it's safe to fetch.\n * Only allows HTTP/HTTPS protocols and validates the URL format.\n *\n * @param urlString - The URL string to validate\n * @throws ValidationError if the URL is invalid or uses an unsafe protocol\n */\nexport function validateUrl(urlString: string): void {\n // Reject URLs with spaces or other invalid characters that shouldn't be in URLs\n if (/\\s/.test(urlString)) {\n throw new ValidationError(\n `Invalid URL: contains whitespace characters`,\n \"url\",\n );\n }\n\n try {\n const url = new URL(urlString);\n\n // Only allow HTTP and HTTPS protocols\n if (url.protocol !== \"http:\" && url.protocol !== \"https:\") {\n throw new ValidationError(\n `Invalid URL protocol: ${url.protocol}. Only http: and https: are allowed.`,\n \"url\",\n );\n }\n\n const hostname = url.hostname.toLowerCase();\n\n // Block localhost (case-insensitive)\n if (hostname === \"localhost\") {\n throw new ValidationError(\n \"Access to localhost addresses is not allowed\",\n \"url\",\n );\n }\n\n // Strip brackets from IPv6 addresses for ipaddr.js\n const cleanHostname = hostname.replace(/^\\[|\\]$/g, \"\");\n\n // Block IP addresses using ipaddr.js\n // This library handles alternative notations (decimal, octal, hex)\n // and comprehensive private range detection for both IPv4 and IPv6\n if (ipaddr.isValid(cleanHostname)) {\n const addr = ipaddr.parse(cleanHostname);\n\n // Check for IPv4 private ranges\n if (addr.kind() === \"ipv4\") {\n const ipv4Addr = addr as ipaddr.IPv4;\n const range = ipv4Addr.range();\n\n if (BLOCKED_IPV4_RANGES.has(range)) {\n throw new ValidationError(\n \"Access to private IP addresses is not allowed\",\n \"url\",\n );\n }\n }\n\n // Check for IPv6 private ranges\n if (addr.kind() === \"ipv6\") {\n const ipv6Addr = addr as ipaddr.IPv6;\n const range = ipv6Addr.range();\n\n if (BLOCKED_IPV6_RANGES.has(range)) {\n throw new ValidationError(\n \"Access to private IP addresses is not allowed\",\n \"url\",\n );\n }\n }\n }\n } catch (error) {\n if (error instanceof ValidationError) {\n throw error;\n }\n throw new ValidationError(\n `Invalid URL format: ${urlString}`,\n \"url\",\n error as Error,\n );\n }\n}\n\n/**\n * Validates a URL and returns the parsed URL object if valid.\n *\n * @param urlString - The URL string to validate\n * @returns The parsed URL object\n * @throws ValidationError if the URL is invalid\n */\nexport function parseValidatedUrl(urlString: string): URL {\n validateUrl(urlString);\n return new URL(urlString);\n}\n"],"mappings":";;;;;;;;;AAUA,MAAM,sBAAsB,IAAI,IAAI;CAClC;CACA;CACA;CACA;CACA;CACA;CACA;CACD,CAAC;AAGF,MAAM,sBAAsB,IAAI,IAAI;CAClC;CACA;CACA;CACA;CACA;CACA;CACA;CACD,CAAC;;;;;;;;AASF,SAAgB,YAAY,WAAyB;AAEnD,KAAI,KAAK,KAAK,UAAU,CACtB,OAAM,IAAIA,8BACR,+CACA,MACD;AAGH,KAAI;EACF,MAAM,MAAM,IAAI,IAAI,UAAU;AAG9B,MAAI,IAAI,aAAa,WAAW,IAAI,aAAa,SAC/C,OAAM,IAAIA,8BACR,yBAAyB,IAAI,SAAS,uCACtC,MACD;EAGH,MAAM,WAAW,IAAI,SAAS,aAAa;AAG3C,MAAI,aAAa,YACf,OAAM,IAAIA,8BACR,gDACA,MACD;EAIH,MAAM,gBAAgB,SAAS,QAAQ,YAAY,GAAG;AAKtD,MAAIC,kBAAO,QAAQ,cAAc,EAAE;GACjC,MAAM,OAAOA,kBAAO,MAAM,cAAc;AAGxC,OAAI,KAAK,MAAM,KAAK,QAAQ;IAE1B,MAAM,QADW,KACM,OAAO;AAE9B,QAAI,oBAAoB,IAAI,MAAM,CAChC,OAAM,IAAID,8BACR,iDACA,MACD;;AAKL,OAAI,KAAK,MAAM,KAAK,QAAQ;IAE1B,MAAM,QADW,KACM,OAAO;AAE9B,QAAI,oBAAoB,IAAI,MAAM,CAChC,OAAM,IAAIA,8BACR,iDACA,MACD;;;UAIA,OAAO;AACd,MAAI,iBAAiBA,8BACnB,OAAM;AAER,QAAM,IAAIA,8BACR,uBAAuB,aACvB,OACA,MACD"}
@@ -0,0 +1,8 @@
1
+ //#region rolldown:runtime
2
+ var __require = /* @__PURE__ */ ((x) => typeof require !== "undefined" ? require : typeof Proxy !== "undefined" ? new Proxy(x, { get: (a, b) => (typeof require !== "undefined" ? require : a)[b] }) : x)(function(x) {
3
+ if (typeof require !== "undefined") return require.apply(this, arguments);
4
+ throw Error("Calling `require` for \"" + x + "\" in an environment that doesn't expose the `require` function.");
5
+ });
6
+
7
+ //#endregion
8
+ export { __require };
@@ -0,0 +1,35 @@
1
+ import { Pinner } from "../../pinner.js";
2
+ import { PinByHashOptions, PinataFile, PinataListBuilder, PinataPin, PinataUploadBuilder, PinataUploadResult } from "../../types/pinata.js";
3
+
4
+ //#region src/adapters/pinata/adapter.d.ts
5
+
6
+ /**
7
+ * Creates a Pinata-compatible adapter for the Pinner client.
8
+ * This allows users to migrate from Pinata SDK with minimal code changes.
9
+ */
10
+ interface PinataAdapter {
11
+ upload: {
12
+ file(file: File): PinataUploadBuilder<PinataUploadResult>;
13
+ fileArray(files: File[]): PinataUploadBuilder<PinataUploadResult>;
14
+ json(data: object): PinataUploadBuilder<PinataUploadResult>;
15
+ base64(base64String: string): PinataUploadBuilder<PinataUploadResult>;
16
+ url(urlString: string): PinataUploadBuilder<PinataUploadResult>;
17
+ cid(cidString: string): PinataUploadBuilder<void>;
18
+ };
19
+ pinByHash(cid: string, options?: PinByHashOptions): Promise<void>;
20
+ unpin(cid: string): Promise<void>;
21
+ getPinStatus(cid: string): Promise<PinataPin>;
22
+ isPinned(cid: string): Promise<boolean>;
23
+ setPinMetadata(cid: string, metadata: Record<string, string>): Promise<void>;
24
+ files: {
25
+ list(): PinataListBuilder<PinataFile[]>;
26
+ get(id: string): Promise<PinataFile>;
27
+ };
28
+ }
29
+ /**
30
+ * Creates a Pinata-compatible adapter for the Pinner client.
31
+ */
32
+ declare function pinataAdapter(pinner: Pinner): PinataAdapter;
33
+ //#endregion
34
+ export { PinataAdapter, pinataAdapter };
35
+ //# sourceMappingURL=adapter.d.ts.map
@@ -0,0 +1,87 @@
1
+ import { createBase64UploadBuilder, createCidUploadBuilder, createFileArrayUploadBuilder, createJsonUploadBuilder, createUploadBuilder, createUrlUploadBuilder } from "./builder.js";
2
+ import { createListBuilder } from "./list-builder.js";
3
+ import { CID } from "multiformats/cid";
4
+
5
+ //#region src/adapters/pinata/adapter.ts
6
+ /**
7
+ * Creates a Pinata-compatible adapter for the Pinner client.
8
+ */
9
+ function pinataAdapter(pinner) {
10
+ return {
11
+ upload: {
12
+ file(file) {
13
+ return createUploadBuilder(pinner, file);
14
+ },
15
+ fileArray(files) {
16
+ return createFileArrayUploadBuilder(pinner, files);
17
+ },
18
+ json(data) {
19
+ return createJsonUploadBuilder(pinner, data);
20
+ },
21
+ base64(base64String) {
22
+ return createBase64UploadBuilder(pinner, base64String);
23
+ },
24
+ url(urlString) {
25
+ return createUrlUploadBuilder(pinner, urlString);
26
+ },
27
+ cid(cidString) {
28
+ return createCidUploadBuilder(pinner, cidString);
29
+ }
30
+ },
31
+ async pinByHash(cid, options) {
32
+ const cidObj = CID.parse(cid);
33
+ const generator = await pinner.pinByHash(cidObj, {
34
+ name: options?.name,
35
+ metadata: options?.keyvalues
36
+ });
37
+ for await (const _ of generator);
38
+ },
39
+ async unpin(cid) {
40
+ await pinner.unpin(cid);
41
+ },
42
+ async getPinStatus(cid) {
43
+ const cidObj = CID.parse(cid);
44
+ const remotePin = await pinner.getPinStatus(cidObj);
45
+ return {
46
+ id: cidObj.toString(),
47
+ ipfsPinHash: cidObj.toString(),
48
+ size: remotePin.size || 0,
49
+ userId: "",
50
+ datePinned: remotePin.created.toISOString(),
51
+ metadata: {
52
+ name: remotePin.name,
53
+ keyvalues: remotePin.metadata
54
+ }
55
+ };
56
+ },
57
+ async isPinned(cid) {
58
+ const cidObj = CID.parse(cid);
59
+ return pinner.isPinned(cidObj);
60
+ },
61
+ async setPinMetadata(cid, metadata) {
62
+ const cidObj = CID.parse(cid);
63
+ await pinner.setPinMetadata(cidObj, metadata);
64
+ },
65
+ files: {
66
+ list() {
67
+ return createListBuilder(pinner);
68
+ },
69
+ async get(id) {
70
+ const pin = (await pinner.listPins()).find((p) => p.cid.toString() === id);
71
+ if (!pin) throw new Error(`Pin not found: ${id}`);
72
+ return {
73
+ id: pin.cid.toString(),
74
+ ipfsPinHash: pin.cid.toString(),
75
+ size: pin.size || 0,
76
+ name: pin.name || "",
77
+ cid: pin.cid.toString(),
78
+ createdAt: pin.created.toISOString()
79
+ };
80
+ }
81
+ }
82
+ };
83
+ }
84
+
85
+ //#endregion
86
+ export { pinataAdapter };
87
+ //# sourceMappingURL=adapter.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"adapter.js","names":[],"sources":["../../../../src/adapters/pinata/adapter.ts"],"sourcesContent":["import type { Pinner } from \"@/pinner\";\nimport {\n createBase64UploadBuilder,\n createCidUploadBuilder,\n createFileArrayUploadBuilder,\n createJsonUploadBuilder,\n createUploadBuilder,\n createUrlUploadBuilder,\n} from \"./builder\";\nimport { createListBuilder } from \"./list-builder\";\nimport type {\n PinataFile,\n PinataListBuilder,\n PinataPin,\n PinataUploadBuilder,\n PinataUploadResult,\n PinByHashOptions,\n} from \"@/types/pinata\";\nimport { CID } from \"multiformats/cid\";\n\n/**\n * Creates a Pinata-compatible adapter for the Pinner client.\n * This allows users to migrate from Pinata SDK with minimal code changes.\n */\nexport interface PinataAdapter {\n upload: {\n file(file: File): PinataUploadBuilder<PinataUploadResult>;\n fileArray(files: File[]): PinataUploadBuilder<PinataUploadResult>;\n json(data: object): PinataUploadBuilder<PinataUploadResult>;\n base64(base64String: string): PinataUploadBuilder<PinataUploadResult>;\n url(urlString: string): PinataUploadBuilder<PinataUploadResult>;\n cid(cidString: string): PinataUploadBuilder<void>;\n };\n\n pinByHash(cid: string, options?: PinByHashOptions): Promise<void>;\n unpin(cid: string): Promise<void>;\n getPinStatus(cid: string): Promise<PinataPin>;\n isPinned(cid: string): Promise<boolean>;\n setPinMetadata(cid: string, metadata: Record<string, string>): Promise<void>;\n\n files: {\n list(): PinataListBuilder<PinataFile[]>;\n get(id: string): Promise<PinataFile>;\n };\n}\n\n/**\n * Creates a Pinata-compatible adapter for the Pinner client.\n */\nexport function pinataAdapter(pinner: Pinner): PinataAdapter {\n return {\n upload: {\n file(file: File): PinataUploadBuilder<PinataUploadResult> {\n return createUploadBuilder(pinner, file);\n },\n fileArray(files: File[]): PinataUploadBuilder<PinataUploadResult> {\n return createFileArrayUploadBuilder(pinner, files);\n },\n json(data: object): PinataUploadBuilder<PinataUploadResult> {\n return createJsonUploadBuilder(pinner, data);\n },\n base64(base64String: string): PinataUploadBuilder<PinataUploadResult> {\n return createBase64UploadBuilder(pinner, base64String);\n },\n url(urlString: string): PinataUploadBuilder<PinataUploadResult> {\n return createUrlUploadBuilder(pinner, urlString);\n },\n cid(cidString: string): PinataUploadBuilder<void> {\n return createCidUploadBuilder(pinner, cidString);\n },\n },\n\n async pinByHash(cid: string, options?: PinByHashOptions): Promise<void> {\n const cidObj = CID.parse(cid);\n const generatorPromise = pinner.pinByHash(cidObj, {\n name: options?.name,\n metadata: options?.keyvalues,\n });\n\n const generator = await generatorPromise;\n for await (const _ of generator) {\n // Pin operation in progress\n }\n },\n\n async unpin(cid: string): Promise<void> {\n await pinner.unpin(cid);\n },\n\n async getPinStatus(cid: string): Promise<PinataPin> {\n const cidObj = CID.parse(cid);\n const remotePin = await pinner.getPinStatus(cidObj);\n\n return {\n id: cidObj.toString(),\n ipfsPinHash: cidObj.toString(),\n size: remotePin.size || 0,\n userId: \"\",\n datePinned: remotePin.created.toISOString(),\n metadata: {\n name: remotePin.name,\n keyvalues: remotePin.metadata,\n },\n };\n },\n\n async isPinned(cid: string): Promise<boolean> {\n const cidObj = CID.parse(cid);\n return pinner.isPinned(cidObj);\n },\n\n async setPinMetadata(\n cid: string,\n metadata: Record<string, string>,\n ): Promise<void> {\n const cidObj = CID.parse(cid);\n await pinner.setPinMetadata(cidObj, metadata);\n },\n\n files: {\n list(): PinataListBuilder<PinataFile[]> {\n return createListBuilder(pinner);\n },\n async get(id: string): Promise<PinataFile> {\n const pins = await pinner.listPins();\n const pin = pins.find((p) => p.cid.toString() === id);\n\n if (!pin) {\n throw new Error(`Pin not found: ${id}`);\n }\n\n return {\n id: pin.cid.toString(),\n ipfsPinHash: pin.cid.toString(),\n size: pin.size || 0,\n name: pin.name || \"\",\n cid: pin.cid.toString(),\n createdAt: pin.created.toISOString(),\n };\n },\n },\n };\n}\n"],"mappings":";;;;;;;;AAiDA,SAAgB,cAAc,QAA+B;AAC3D,QAAO;EACL,QAAQ;GACN,KAAK,MAAqD;AACxD,WAAO,oBAAoB,QAAQ,KAAK;;GAE1C,UAAU,OAAwD;AAChE,WAAO,6BAA6B,QAAQ,MAAM;;GAEpD,KAAK,MAAuD;AAC1D,WAAO,wBAAwB,QAAQ,KAAK;;GAE9C,OAAO,cAA+D;AACpE,WAAO,0BAA0B,QAAQ,aAAa;;GAExD,IAAI,WAA4D;AAC9D,WAAO,uBAAuB,QAAQ,UAAU;;GAElD,IAAI,WAA8C;AAChD,WAAO,uBAAuB,QAAQ,UAAU;;GAEnD;EAED,MAAM,UAAU,KAAa,SAA2C;GACtE,MAAM,SAAS,IAAI,MAAM,IAAI;GAM7B,MAAM,YAAY,MALO,OAAO,UAAU,QAAQ;IAChD,MAAM,SAAS;IACf,UAAU,SAAS;IACpB,CAAC;AAGF,cAAW,MAAM,KAAK;;EAKxB,MAAM,MAAM,KAA4B;AACtC,SAAM,OAAO,MAAM,IAAI;;EAGzB,MAAM,aAAa,KAAiC;GAClD,MAAM,SAAS,IAAI,MAAM,IAAI;GAC7B,MAAM,YAAY,MAAM,OAAO,aAAa,OAAO;AAEnD,UAAO;IACL,IAAI,OAAO,UAAU;IACrB,aAAa,OAAO,UAAU;IAC9B,MAAM,UAAU,QAAQ;IACxB,QAAQ;IACR,YAAY,UAAU,QAAQ,aAAa;IAC3C,UAAU;KACR,MAAM,UAAU;KAChB,WAAW,UAAU;KACtB;IACF;;EAGH,MAAM,SAAS,KAA+B;GAC5C,MAAM,SAAS,IAAI,MAAM,IAAI;AAC7B,UAAO,OAAO,SAAS,OAAO;;EAGhC,MAAM,eACJ,KACA,UACe;GACf,MAAM,SAAS,IAAI,MAAM,IAAI;AAC7B,SAAM,OAAO,eAAe,QAAQ,SAAS;;EAG/C,OAAO;GACL,OAAwC;AACtC,WAAO,kBAAkB,OAAO;;GAElC,MAAM,IAAI,IAAiC;IAEzC,MAAM,OADO,MAAM,OAAO,UAAU,EACnB,MAAM,MAAM,EAAE,IAAI,UAAU,KAAK,GAAG;AAErD,QAAI,CAAC,IACH,OAAM,IAAI,MAAM,kBAAkB,KAAK;AAGzC,WAAO;KACL,IAAI,IAAI,IAAI,UAAU;KACtB,aAAa,IAAI,IAAI,UAAU;KAC/B,MAAM,IAAI,QAAQ;KAClB,MAAM,IAAI,QAAQ;KAClB,KAAK,IAAI,IAAI,UAAU;KACvB,WAAW,IAAI,QAAQ,aAAa;KACrC;;GAEJ;EACF"}
@@ -0,0 +1 @@
1
+ import "../../pinner.js";