@elqnt/react 1.0.2 → 1.0.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/components/form-controls/icon.js +54 -3
- package/dist/components/form-controls/icon.js.map +1 -1
- package/dist/components/form-controls/icon.mjs +28 -3
- package/dist/components/form-controls/icon.mjs.map +1 -1
- package/dist/components/upload/upload-actions.js +147 -8
- package/dist/components/upload/upload-actions.js.map +1 -1
- package/dist/components/upload/upload-actions.mjs +118 -4
- package/dist/components/upload/upload-actions.mjs.map +1 -1
- package/dist/index.js +1414 -1201
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +626 -483
- package/dist/index.mjs.map +1 -1
- package/package.json +1 -1
- package/dist/chunk-FLTEIMWJ.mjs +0 -36
- package/dist/chunk-FLTEIMWJ.mjs.map +0 -1
- package/dist/chunk-UXJQ7CBV.mjs +0 -122
- package/dist/chunk-UXJQ7CBV.mjs.map +0 -1
- package/dist/chunk-W4AIK44N.js +0 -122
- package/dist/chunk-W4AIK44N.js.map +0 -1
- package/dist/chunk-ZCBGKBZS.js +0 -36
- package/dist/chunk-ZCBGKBZS.js.map +0 -1
package/package.json
CHANGED
package/dist/chunk-FLTEIMWJ.mjs
DELETED
|
@@ -1,36 +0,0 @@
|
|
|
1
|
-
// lib/utils.ts
|
|
2
|
-
import { clsx } from "clsx";
|
|
3
|
-
import { twMerge } from "tailwind-merge";
|
|
4
|
-
function cn(...inputs) {
|
|
5
|
-
return twMerge(clsx(inputs));
|
|
6
|
-
}
|
|
7
|
-
function getBadgeVariant(status) {
|
|
8
|
-
return status === "open" ? "default" : "secondary";
|
|
9
|
-
}
|
|
10
|
-
|
|
11
|
-
// components/form-controls/icon.tsx
|
|
12
|
-
import { icons } from "lucide-react";
|
|
13
|
-
import { memo } from "react";
|
|
14
|
-
import { jsx } from "react/jsx-runtime";
|
|
15
|
-
var Icon = memo(({ name, className, strokeWidth, size }) => {
|
|
16
|
-
const IconComponent = icons[name];
|
|
17
|
-
if (!IconComponent) {
|
|
18
|
-
return null;
|
|
19
|
-
}
|
|
20
|
-
return /* @__PURE__ */ jsx(
|
|
21
|
-
IconComponent,
|
|
22
|
-
{
|
|
23
|
-
className: cn(className),
|
|
24
|
-
strokeWidth: strokeWidth || 2.5,
|
|
25
|
-
size
|
|
26
|
-
}
|
|
27
|
-
);
|
|
28
|
-
});
|
|
29
|
-
Icon.displayName = "Icon";
|
|
30
|
-
|
|
31
|
-
export {
|
|
32
|
-
cn,
|
|
33
|
-
getBadgeVariant,
|
|
34
|
-
Icon
|
|
35
|
-
};
|
|
36
|
-
//# sourceMappingURL=chunk-FLTEIMWJ.mjs.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../lib/utils.ts","../components/form-controls/icon.tsx"],"sourcesContent":["import { type ClassValue, clsx } from \"clsx\";\nimport { twMerge } from \"tailwind-merge\";\n\nexport function cn(...inputs: ClassValue[]) {\n return twMerge(clsx(inputs));\n}\n\nexport function getBadgeVariant(status: string) {\n return status === \"open\" ? \"default\" : \"secondary\";\n}\n\n// export function getColspanByFieldType(fieldType: EntityFieldType) {\n// return fieldType === \"text\" ? \"col-span-2\" : \"\";\n// }\n","import { cn } from \"../../lib/utils\";\nimport { icons } from \"lucide-react\";\nimport { memo } from \"react\";\n\nexport type IconProps = {\n name: keyof typeof icons;\n size?: number;\n className?: string;\n strokeWidth?: number;\n};\n\nexport const Icon = memo(({ name, className, strokeWidth, size }: IconProps) => {\n const IconComponent = icons[name];\n\n if (!IconComponent) {\n return null;\n }\n\n return (\n <IconComponent\n className={cn(className)}\n strokeWidth={strokeWidth || 2.5}\n size={size}\n />\n );\n});\n\nIcon.displayName = \"Icon\";\n"],"mappings":";AAAA,SAA0B,YAAY;AACtC,SAAS,eAAe;AAEjB,SAAS,MAAM,QAAsB;AAC1C,SAAO,QAAQ,KAAK,MAAM,CAAC;AAC7B;AAEO,SAAS,gBAAgB,QAAgB;AAC9C,SAAO,WAAW,SAAS,YAAY;AACzC;;;ACRA,SAAS,aAAa;AACtB,SAAS,YAAY;AAiBjB;AARG,IAAM,OAAO,KAAK,CAAC,EAAE,MAAM,WAAW,aAAa,KAAK,MAAiB;AAC9E,QAAM,gBAAgB,MAAM,IAAI;AAEhC,MAAI,CAAC,eAAe;AAClB,WAAO;AAAA,EACT;AAEA,SACE;AAAA,IAAC;AAAA;AAAA,MACC,WAAW,GAAG,SAAS;AAAA,MACvB,aAAa,eAAe;AAAA,MAC5B;AAAA;AAAA,EACF;AAEJ,CAAC;AAED,KAAK,cAAc;","names":[]}
|
package/dist/chunk-UXJQ7CBV.mjs
DELETED
|
@@ -1,122 +0,0 @@
|
|
|
1
|
-
// components/upload/upload-actions.ts
|
|
2
|
-
import { BlobServiceClient } from "@azure/storage-blob";
|
|
3
|
-
import { PutObjectCommand, S3Client } from "@aws-sdk/client-s3";
|
|
4
|
-
async function uploadFile(formData) {
|
|
5
|
-
try {
|
|
6
|
-
const file = formData.get("file");
|
|
7
|
-
if (!file) {
|
|
8
|
-
throw new Error("No file provided");
|
|
9
|
-
}
|
|
10
|
-
const STORAGE_CONNECTION_STRING = process.env.AZURE_STORAGE_CONNECTION_STRING;
|
|
11
|
-
const STORAGE_CONTAINER_NAME = process.env.AZURE_STORAGE_CONTAINER_NAME;
|
|
12
|
-
if (!STORAGE_CONNECTION_STRING || !STORAGE_CONTAINER_NAME) {
|
|
13
|
-
throw new Error(
|
|
14
|
-
"STORAGE_CONNECTION_STRING or STORAGE_CONTAINER_NAME not set"
|
|
15
|
-
);
|
|
16
|
-
}
|
|
17
|
-
const fileExtension = file.name.split(".").pop()?.toLowerCase();
|
|
18
|
-
const fileName = `${Date.now()}-${Math.random().toString(36).substring(2)}.${fileExtension}`;
|
|
19
|
-
const contentType = getContentType(fileExtension);
|
|
20
|
-
const buffer = Buffer.from(await file.arrayBuffer());
|
|
21
|
-
const blobServiceClient = BlobServiceClient.fromConnectionString(
|
|
22
|
-
STORAGE_CONNECTION_STRING
|
|
23
|
-
);
|
|
24
|
-
const containerClient = blobServiceClient.getContainerClient(
|
|
25
|
-
STORAGE_CONTAINER_NAME
|
|
26
|
-
);
|
|
27
|
-
await containerClient.createIfNotExists();
|
|
28
|
-
const blockBlobClient = containerClient.getBlockBlobClient(fileName);
|
|
29
|
-
await blockBlobClient.upload(buffer, buffer.length, {
|
|
30
|
-
blobHTTPHeaders: {
|
|
31
|
-
blobContentType: contentType
|
|
32
|
-
}
|
|
33
|
-
});
|
|
34
|
-
const fileUrl = blockBlobClient.url;
|
|
35
|
-
return { success: true, fileUrl };
|
|
36
|
-
} catch (error) {
|
|
37
|
-
console.error("Upload error:", error);
|
|
38
|
-
return {
|
|
39
|
-
success: false,
|
|
40
|
-
error: error instanceof Error ? error.message : "Failed to upload file"
|
|
41
|
-
};
|
|
42
|
-
}
|
|
43
|
-
}
|
|
44
|
-
function getContentType(fileExtension) {
|
|
45
|
-
if (!fileExtension) return "application/octet-stream";
|
|
46
|
-
const mimeTypes = {
|
|
47
|
-
// Images
|
|
48
|
-
png: "image/png",
|
|
49
|
-
jpg: "image/jpeg",
|
|
50
|
-
jpeg: "image/jpeg",
|
|
51
|
-
gif: "image/gif",
|
|
52
|
-
webp: "image/webp",
|
|
53
|
-
svg: "image/svg+xml",
|
|
54
|
-
// Documents
|
|
55
|
-
pdf: "application/pdf",
|
|
56
|
-
doc: "application/msword",
|
|
57
|
-
docx: "application/vnd.openxmlformats-officedocument.wordprocessingml.document",
|
|
58
|
-
// Others
|
|
59
|
-
txt: "text/plain",
|
|
60
|
-
csv: "text/csv",
|
|
61
|
-
json: "application/json"
|
|
62
|
-
};
|
|
63
|
-
return mimeTypes[fileExtension] || "application/octet-stream";
|
|
64
|
-
}
|
|
65
|
-
async function uploadToS3(formData) {
|
|
66
|
-
try {
|
|
67
|
-
const file = formData.get("file");
|
|
68
|
-
if (!file) {
|
|
69
|
-
throw new Error("No file provided");
|
|
70
|
-
}
|
|
71
|
-
const AWS_ACCESS_KEY_ID = process.env.LINODE_ACCESS_KEY;
|
|
72
|
-
const AWS_SECRET_ACCESS_KEY = process.env.LINODE_SECRET_KEY;
|
|
73
|
-
const S3_BUCKET_NAME = process.env.LINODE_BUCKET_NAME;
|
|
74
|
-
let S3_ENDPOINT = process.env.LINODE_OBJECT_STORAGE_ENDPOINT;
|
|
75
|
-
if (!AWS_ACCESS_KEY_ID || !AWS_SECRET_ACCESS_KEY || !S3_BUCKET_NAME || !S3_ENDPOINT) {
|
|
76
|
-
throw new Error("S3 credentials or configuration not set");
|
|
77
|
-
}
|
|
78
|
-
const fileExtension = file.name.split(".").pop()?.toLowerCase();
|
|
79
|
-
const fileName = `${Date.now()}-${Math.random().toString(36).substring(2)}.${fileExtension}`;
|
|
80
|
-
const contentType = getContentType(fileExtension);
|
|
81
|
-
const buffer = Buffer.from(await file.arrayBuffer());
|
|
82
|
-
const s3Client = new S3Client({
|
|
83
|
-
endpoint: S3_ENDPOINT,
|
|
84
|
-
region: "us-east-1",
|
|
85
|
-
// Linode requires a region but it's not used with custom endpoint
|
|
86
|
-
credentials: {
|
|
87
|
-
accessKeyId: AWS_ACCESS_KEY_ID,
|
|
88
|
-
secretAccessKey: AWS_SECRET_ACCESS_KEY
|
|
89
|
-
},
|
|
90
|
-
forcePathStyle: true,
|
|
91
|
-
// Required for Linode Object Storage
|
|
92
|
-
maxAttempts: 3,
|
|
93
|
-
requestHandler: {
|
|
94
|
-
timeout: 1e4
|
|
95
|
-
// 10 seconds timeout
|
|
96
|
-
}
|
|
97
|
-
});
|
|
98
|
-
const command = new PutObjectCommand({
|
|
99
|
-
Bucket: S3_BUCKET_NAME,
|
|
100
|
-
Key: fileName,
|
|
101
|
-
Body: buffer,
|
|
102
|
-
ContentType: contentType,
|
|
103
|
-
ACL: "public-read"
|
|
104
|
-
// Make the object publicly readable
|
|
105
|
-
});
|
|
106
|
-
await s3Client.send(command);
|
|
107
|
-
const fileUrl = `${S3_ENDPOINT}/${S3_BUCKET_NAME}/${fileName}`;
|
|
108
|
-
return { success: true, fileUrl };
|
|
109
|
-
} catch (error) {
|
|
110
|
-
console.error("Upload error:", JSON.stringify(error, null, 2));
|
|
111
|
-
return {
|
|
112
|
-
success: false,
|
|
113
|
-
error: error instanceof Error ? error.message : "Failed to upload file"
|
|
114
|
-
};
|
|
115
|
-
}
|
|
116
|
-
}
|
|
117
|
-
|
|
118
|
-
export {
|
|
119
|
-
uploadFile,
|
|
120
|
-
uploadToS3
|
|
121
|
-
};
|
|
122
|
-
//# sourceMappingURL=chunk-UXJQ7CBV.mjs.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["../components/upload/upload-actions.ts"],"sourcesContent":["\"use server\";\nimport { BlobServiceClient } from \"@azure/storage-blob\";\nimport { PutObjectCommand, S3Client } from \"@aws-sdk/client-s3\";\n\n\nexport async function uploadFile(formData: FormData) {\n try {\n const file = formData.get(\"file\") as File;\n if (!file) {\n throw new Error(\"No file provided\");\n }\n\n const STORAGE_CONNECTION_STRING =\n process.env.AZURE_STORAGE_CONNECTION_STRING;\n const STORAGE_CONTAINER_NAME = process.env.AZURE_STORAGE_CONTAINER_NAME;\n\n if (!STORAGE_CONNECTION_STRING || !STORAGE_CONTAINER_NAME) {\n throw new Error(\n \"STORAGE_CONNECTION_STRING or STORAGE_CONTAINER_NAME not set\"\n );\n }\n\n // Generate unique filename\n const fileExtension = file.name.split(\".\").pop()?.toLowerCase();\n const fileName = `${Date.now()}-${Math.random().toString(36).substring(2)}.${fileExtension}`;\n\n // Determine content type based on file extension\n const contentType = getContentType(fileExtension);\n\n // Convert file to buffer\n const buffer = Buffer.from(await file.arrayBuffer());\n\n // Upload to Azure Blob Storage\n const blobServiceClient = BlobServiceClient.fromConnectionString(\n STORAGE_CONNECTION_STRING\n );\n const containerClient = blobServiceClient.getContainerClient(\n STORAGE_CONTAINER_NAME\n );\n\n // Create container if it doesn't exist\n await containerClient.createIfNotExists();\n\n // Upload file with content type\n const blockBlobClient = containerClient.getBlockBlobClient(fileName);\n await blockBlobClient.upload(buffer, buffer.length, {\n blobHTTPHeaders: {\n blobContentType: contentType,\n },\n });\n\n // Get the URL\n const fileUrl = blockBlobClient.url;\n\n return { success: true, fileUrl };\n } catch (error) {\n console.error(\"Upload error:\", error);\n return {\n success: false,\n error: error instanceof Error ? error.message : \"Failed to upload file\",\n };\n }\n}\n\nfunction getContentType(fileExtension: string | undefined): string {\n if (!fileExtension) return \"application/octet-stream\";\n\n const mimeTypes: Record<string, string> = {\n // Images\n png: \"image/png\",\n jpg: \"image/jpeg\",\n jpeg: \"image/jpeg\",\n gif: \"image/gif\",\n webp: \"image/webp\",\n svg: \"image/svg+xml\",\n // Documents\n pdf: \"application/pdf\",\n doc: \"application/msword\",\n docx: \"application/vnd.openxmlformats-officedocument.wordprocessingml.document\",\n // Others\n txt: \"text/plain\",\n csv: \"text/csv\",\n json: \"application/json\",\n };\n\n return mimeTypes[fileExtension] || \"application/octet-stream\";\n}\n\nexport async function uploadToS3(formData: FormData) {\n try {\n const file = formData.get(\"file\") as File;\n if (!file) {\n throw new Error(\"No file provided\");\n }\n\n const AWS_ACCESS_KEY_ID = process.env.LINODE_ACCESS_KEY;\n const AWS_SECRET_ACCESS_KEY = process.env.LINODE_SECRET_KEY;\n const S3_BUCKET_NAME = process.env.LINODE_BUCKET_NAME;\n let S3_ENDPOINT = process.env.LINODE_OBJECT_STORAGE_ENDPOINT;\n\n if (\n !AWS_ACCESS_KEY_ID ||\n !AWS_SECRET_ACCESS_KEY ||\n !S3_BUCKET_NAME ||\n !S3_ENDPOINT\n ) {\n throw new Error(\"S3 credentials or configuration not set\");\n }\n\n // Generate unique filename\n const fileExtension = file.name.split(\".\").pop()?.toLowerCase();\n const fileName = `${Date.now()}-${Math.random().toString(36).substring(2)}.${fileExtension}`;\n\n // Determine content type based on file extension\n const contentType = getContentType(fileExtension);\n\n // Convert file to buffer\n const buffer = Buffer.from(await file.arrayBuffer());\n\n // Initialize S3 client\n const s3Client = new S3Client({\n endpoint: S3_ENDPOINT,\n region: \"us-east-1\", // Linode requires a region but it's not used with custom endpoint\n credentials: {\n accessKeyId: AWS_ACCESS_KEY_ID,\n secretAccessKey: AWS_SECRET_ACCESS_KEY,\n },\n forcePathStyle: true, // Required for Linode Object Storage\n maxAttempts: 3,\n requestHandler: {\n timeout: 10000, // 10 seconds timeout\n },\n });\n\n // Upload to S3\n const command = new PutObjectCommand({\n Bucket: S3_BUCKET_NAME,\n Key: fileName,\n Body: buffer,\n ContentType: contentType,\n ACL: \"public-read\", // Make the object publicly readable\n });\n\n await s3Client.send(command);\n\n // Construct the URL using the endpoint\n const fileUrl = `${S3_ENDPOINT}/${S3_BUCKET_NAME}/${fileName}`;\n\n return { success: true, fileUrl };\n } catch (error) {\n console.error(\"Upload error:\", JSON.stringify(error, null, 2));\n return {\n success: false,\n error: error instanceof Error ? error.message : \"Failed to upload file\",\n };\n }\n}\n"],"mappings":";AACA,SAAS,yBAAyB;AAClC,SAAS,kBAAkB,gBAAgB;AAG3C,eAAsB,WAAW,UAAoB;AACnD,MAAI;AACF,UAAM,OAAO,SAAS,IAAI,MAAM;AAChC,QAAI,CAAC,MAAM;AACT,YAAM,IAAI,MAAM,kBAAkB;AAAA,IACpC;AAEA,UAAM,4BACJ,QAAQ,IAAI;AACd,UAAM,yBAAyB,QAAQ,IAAI;AAE3C,QAAI,CAAC,6BAA6B,CAAC,wBAAwB;AACzD,YAAM,IAAI;AAAA,QACR;AAAA,MACF;AAAA,IACF;AAGA,UAAM,gBAAgB,KAAK,KAAK,MAAM,GAAG,EAAE,IAAI,GAAG,YAAY;AAC9D,UAAM,WAAW,GAAG,KAAK,IAAI,CAAC,IAAI,KAAK,OAAO,EAAE,SAAS,EAAE,EAAE,UAAU,CAAC,CAAC,IAAI,aAAa;AAG1F,UAAM,cAAc,eAAe,aAAa;AAGhD,UAAM,SAAS,OAAO,KAAK,MAAM,KAAK,YAAY,CAAC;AAGnD,UAAM,oBAAoB,kBAAkB;AAAA,MAC1C;AAAA,IACF;AACA,UAAM,kBAAkB,kBAAkB;AAAA,MACxC;AAAA,IACF;AAGA,UAAM,gBAAgB,kBAAkB;AAGxC,UAAM,kBAAkB,gBAAgB,mBAAmB,QAAQ;AACnE,UAAM,gBAAgB,OAAO,QAAQ,OAAO,QAAQ;AAAA,MAClD,iBAAiB;AAAA,QACf,iBAAiB;AAAA,MACnB;AAAA,IACF,CAAC;AAGD,UAAM,UAAU,gBAAgB;AAEhC,WAAO,EAAE,SAAS,MAAM,QAAQ;AAAA,EAClC,SAAS,OAAO;AACd,YAAQ,MAAM,iBAAiB,KAAK;AACpC,WAAO;AAAA,MACL,SAAS;AAAA,MACT,OAAO,iBAAiB,QAAQ,MAAM,UAAU;AAAA,IAClD;AAAA,EACF;AACF;AAEA,SAAS,eAAe,eAA2C;AACjE,MAAI,CAAC,cAAe,QAAO;AAE3B,QAAM,YAAoC;AAAA;AAAA,IAExC,KAAK;AAAA,IACL,KAAK;AAAA,IACL,MAAM;AAAA,IACN,KAAK;AAAA,IACL,MAAM;AAAA,IACN,KAAK;AAAA;AAAA,IAEL,KAAK;AAAA,IACL,KAAK;AAAA,IACL,MAAM;AAAA;AAAA,IAEN,KAAK;AAAA,IACL,KAAK;AAAA,IACL,MAAM;AAAA,EACR;AAEA,SAAO,UAAU,aAAa,KAAK;AACrC;AAEA,eAAsB,WAAW,UAAoB;AACnD,MAAI;AACF,UAAM,OAAO,SAAS,IAAI,MAAM;AAChC,QAAI,CAAC,MAAM;AACT,YAAM,IAAI,MAAM,kBAAkB;AAAA,IACpC;AAEA,UAAM,oBAAoB,QAAQ,IAAI;AACtC,UAAM,wBAAwB,QAAQ,IAAI;AAC1C,UAAM,iBAAiB,QAAQ,IAAI;AACnC,QAAI,cAAc,QAAQ,IAAI;AAE9B,QACE,CAAC,qBACD,CAAC,yBACD,CAAC,kBACD,CAAC,aACD;AACA,YAAM,IAAI,MAAM,yCAAyC;AAAA,IAC3D;AAGA,UAAM,gBAAgB,KAAK,KAAK,MAAM,GAAG,EAAE,IAAI,GAAG,YAAY;AAC9D,UAAM,WAAW,GAAG,KAAK,IAAI,CAAC,IAAI,KAAK,OAAO,EAAE,SAAS,EAAE,EAAE,UAAU,CAAC,CAAC,IAAI,aAAa;AAG1F,UAAM,cAAc,eAAe,aAAa;AAGhD,UAAM,SAAS,OAAO,KAAK,MAAM,KAAK,YAAY,CAAC;AAGnD,UAAM,WAAW,IAAI,SAAS;AAAA,MAC5B,UAAU;AAAA,MACV,QAAQ;AAAA;AAAA,MACR,aAAa;AAAA,QACX,aAAa;AAAA,QACb,iBAAiB;AAAA,MACnB;AAAA,MACA,gBAAgB;AAAA;AAAA,MAChB,aAAa;AAAA,MACb,gBAAgB;AAAA,QACd,SAAS;AAAA;AAAA,MACX;AAAA,IACF,CAAC;AAGD,UAAM,UAAU,IAAI,iBAAiB;AAAA,MACnC,QAAQ;AAAA,MACR,KAAK;AAAA,MACL,MAAM;AAAA,MACN,aAAa;AAAA,MACb,KAAK;AAAA;AAAA,IACP,CAAC;AAED,UAAM,SAAS,KAAK,OAAO;AAG3B,UAAM,UAAU,GAAG,WAAW,IAAI,cAAc,IAAI,QAAQ;AAE5D,WAAO,EAAE,SAAS,MAAM,QAAQ;AAAA,EAClC,SAAS,OAAO;AACd,YAAQ,MAAM,iBAAiB,KAAK,UAAU,OAAO,MAAM,CAAC,CAAC;AAC7D,WAAO;AAAA,MACL,SAAS;AAAA,MACT,OAAO,iBAAiB,QAAQ,MAAM,UAAU;AAAA,IAClD;AAAA,EACF;AACF;","names":[]}
|
package/dist/chunk-W4AIK44N.js
DELETED
|
@@ -1,122 +0,0 @@
|
|
|
1
|
-
"use strict";Object.defineProperty(exports, "__esModule", {value: true}); function _optionalChain(ops) { let lastAccessLHS = undefined; let value = ops[0]; let i = 1; while (i < ops.length) { const op = ops[i]; const fn = ops[i + 1]; i += 2; if ((op === 'optionalAccess' || op === 'optionalCall') && value == null) { return undefined; } if (op === 'access' || op === 'optionalAccess') { lastAccessLHS = value; value = fn(value); } else if (op === 'call' || op === 'optionalCall') { value = fn((...args) => value.call(lastAccessLHS, ...args)); lastAccessLHS = undefined; } } return value; }// components/upload/upload-actions.ts
|
|
2
|
-
var _storageblob = require('@azure/storage-blob');
|
|
3
|
-
var _clients3 = require('@aws-sdk/client-s3');
|
|
4
|
-
async function uploadFile(formData) {
|
|
5
|
-
try {
|
|
6
|
-
const file = formData.get("file");
|
|
7
|
-
if (!file) {
|
|
8
|
-
throw new Error("No file provided");
|
|
9
|
-
}
|
|
10
|
-
const STORAGE_CONNECTION_STRING = process.env.AZURE_STORAGE_CONNECTION_STRING;
|
|
11
|
-
const STORAGE_CONTAINER_NAME = process.env.AZURE_STORAGE_CONTAINER_NAME;
|
|
12
|
-
if (!STORAGE_CONNECTION_STRING || !STORAGE_CONTAINER_NAME) {
|
|
13
|
-
throw new Error(
|
|
14
|
-
"STORAGE_CONNECTION_STRING or STORAGE_CONTAINER_NAME not set"
|
|
15
|
-
);
|
|
16
|
-
}
|
|
17
|
-
const fileExtension = _optionalChain([file, 'access', _ => _.name, 'access', _2 => _2.split, 'call', _3 => _3("."), 'access', _4 => _4.pop, 'call', _5 => _5(), 'optionalAccess', _6 => _6.toLowerCase, 'call', _7 => _7()]);
|
|
18
|
-
const fileName = `${Date.now()}-${Math.random().toString(36).substring(2)}.${fileExtension}`;
|
|
19
|
-
const contentType = getContentType(fileExtension);
|
|
20
|
-
const buffer = Buffer.from(await file.arrayBuffer());
|
|
21
|
-
const blobServiceClient = _storageblob.BlobServiceClient.fromConnectionString(
|
|
22
|
-
STORAGE_CONNECTION_STRING
|
|
23
|
-
);
|
|
24
|
-
const containerClient = blobServiceClient.getContainerClient(
|
|
25
|
-
STORAGE_CONTAINER_NAME
|
|
26
|
-
);
|
|
27
|
-
await containerClient.createIfNotExists();
|
|
28
|
-
const blockBlobClient = containerClient.getBlockBlobClient(fileName);
|
|
29
|
-
await blockBlobClient.upload(buffer, buffer.length, {
|
|
30
|
-
blobHTTPHeaders: {
|
|
31
|
-
blobContentType: contentType
|
|
32
|
-
}
|
|
33
|
-
});
|
|
34
|
-
const fileUrl = blockBlobClient.url;
|
|
35
|
-
return { success: true, fileUrl };
|
|
36
|
-
} catch (error) {
|
|
37
|
-
console.error("Upload error:", error);
|
|
38
|
-
return {
|
|
39
|
-
success: false,
|
|
40
|
-
error: error instanceof Error ? error.message : "Failed to upload file"
|
|
41
|
-
};
|
|
42
|
-
}
|
|
43
|
-
}
|
|
44
|
-
function getContentType(fileExtension) {
|
|
45
|
-
if (!fileExtension) return "application/octet-stream";
|
|
46
|
-
const mimeTypes = {
|
|
47
|
-
// Images
|
|
48
|
-
png: "image/png",
|
|
49
|
-
jpg: "image/jpeg",
|
|
50
|
-
jpeg: "image/jpeg",
|
|
51
|
-
gif: "image/gif",
|
|
52
|
-
webp: "image/webp",
|
|
53
|
-
svg: "image/svg+xml",
|
|
54
|
-
// Documents
|
|
55
|
-
pdf: "application/pdf",
|
|
56
|
-
doc: "application/msword",
|
|
57
|
-
docx: "application/vnd.openxmlformats-officedocument.wordprocessingml.document",
|
|
58
|
-
// Others
|
|
59
|
-
txt: "text/plain",
|
|
60
|
-
csv: "text/csv",
|
|
61
|
-
json: "application/json"
|
|
62
|
-
};
|
|
63
|
-
return mimeTypes[fileExtension] || "application/octet-stream";
|
|
64
|
-
}
|
|
65
|
-
async function uploadToS3(formData) {
|
|
66
|
-
try {
|
|
67
|
-
const file = formData.get("file");
|
|
68
|
-
if (!file) {
|
|
69
|
-
throw new Error("No file provided");
|
|
70
|
-
}
|
|
71
|
-
const AWS_ACCESS_KEY_ID = process.env.LINODE_ACCESS_KEY;
|
|
72
|
-
const AWS_SECRET_ACCESS_KEY = process.env.LINODE_SECRET_KEY;
|
|
73
|
-
const S3_BUCKET_NAME = process.env.LINODE_BUCKET_NAME;
|
|
74
|
-
let S3_ENDPOINT = process.env.LINODE_OBJECT_STORAGE_ENDPOINT;
|
|
75
|
-
if (!AWS_ACCESS_KEY_ID || !AWS_SECRET_ACCESS_KEY || !S3_BUCKET_NAME || !S3_ENDPOINT) {
|
|
76
|
-
throw new Error("S3 credentials or configuration not set");
|
|
77
|
-
}
|
|
78
|
-
const fileExtension = _optionalChain([file, 'access', _8 => _8.name, 'access', _9 => _9.split, 'call', _10 => _10("."), 'access', _11 => _11.pop, 'call', _12 => _12(), 'optionalAccess', _13 => _13.toLowerCase, 'call', _14 => _14()]);
|
|
79
|
-
const fileName = `${Date.now()}-${Math.random().toString(36).substring(2)}.${fileExtension}`;
|
|
80
|
-
const contentType = getContentType(fileExtension);
|
|
81
|
-
const buffer = Buffer.from(await file.arrayBuffer());
|
|
82
|
-
const s3Client = new (0, _clients3.S3Client)({
|
|
83
|
-
endpoint: S3_ENDPOINT,
|
|
84
|
-
region: "us-east-1",
|
|
85
|
-
// Linode requires a region but it's not used with custom endpoint
|
|
86
|
-
credentials: {
|
|
87
|
-
accessKeyId: AWS_ACCESS_KEY_ID,
|
|
88
|
-
secretAccessKey: AWS_SECRET_ACCESS_KEY
|
|
89
|
-
},
|
|
90
|
-
forcePathStyle: true,
|
|
91
|
-
// Required for Linode Object Storage
|
|
92
|
-
maxAttempts: 3,
|
|
93
|
-
requestHandler: {
|
|
94
|
-
timeout: 1e4
|
|
95
|
-
// 10 seconds timeout
|
|
96
|
-
}
|
|
97
|
-
});
|
|
98
|
-
const command = new (0, _clients3.PutObjectCommand)({
|
|
99
|
-
Bucket: S3_BUCKET_NAME,
|
|
100
|
-
Key: fileName,
|
|
101
|
-
Body: buffer,
|
|
102
|
-
ContentType: contentType,
|
|
103
|
-
ACL: "public-read"
|
|
104
|
-
// Make the object publicly readable
|
|
105
|
-
});
|
|
106
|
-
await s3Client.send(command);
|
|
107
|
-
const fileUrl = `${S3_ENDPOINT}/${S3_BUCKET_NAME}/${fileName}`;
|
|
108
|
-
return { success: true, fileUrl };
|
|
109
|
-
} catch (error) {
|
|
110
|
-
console.error("Upload error:", JSON.stringify(error, null, 2));
|
|
111
|
-
return {
|
|
112
|
-
success: false,
|
|
113
|
-
error: error instanceof Error ? error.message : "Failed to upload file"
|
|
114
|
-
};
|
|
115
|
-
}
|
|
116
|
-
}
|
|
117
|
-
|
|
118
|
-
|
|
119
|
-
|
|
120
|
-
|
|
121
|
-
exports.uploadFile = uploadFile; exports.uploadToS3 = uploadToS3;
|
|
122
|
-
//# sourceMappingURL=chunk-W4AIK44N.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["/home/runner/work/eloquent-packages/eloquent-packages/packages/react/dist/chunk-W4AIK44N.js","../components/upload/upload-actions.ts"],"names":[],"mappings":"AAAA;ACCA,kDAAkC;AAClC,8CAA2C;AAG3C,MAAA,SAAsB,UAAA,CAAW,QAAA,EAAoB;AACnD,EAAA,IAAI;AACF,IAAA,MAAM,KAAA,EAAO,QAAA,CAAS,GAAA,CAAI,MAAM,CAAA;AAChC,IAAA,GAAA,CAAI,CAAC,IAAA,EAAM;AACT,MAAA,MAAM,IAAI,KAAA,CAAM,kBAAkB,CAAA;AAAA,IACpC;AAEA,IAAA,MAAM,0BAAA,EACJ,OAAA,CAAQ,GAAA,CAAI,+BAAA;AACd,IAAA,MAAM,uBAAA,EAAyB,OAAA,CAAQ,GAAA,CAAI,4BAAA;AAE3C,IAAA,GAAA,CAAI,CAAC,0BAAA,GAA6B,CAAC,sBAAA,EAAwB;AACzD,MAAA,MAAM,IAAI,KAAA;AAAA,QACR;AAAA,MACF,CAAA;AAAA,IACF;AAGA,IAAA,MAAM,cAAA,kBAAgB,IAAA,mBAAK,IAAA,qBAAK,KAAA,mBAAM,GAAG,CAAA,qBAAE,GAAA,mBAAI,CAAA,6BAAG,WAAA,mBAAY,GAAA;AAC9D,IAAA,MAAM,SAAA,EAAW,CAAA,EAAA;AAGX,IAAA;AAGS,IAAA;AAGT,IAAA;AACJ,MAAA;AACF,IAAA;AACM,IAAA;AACJ,MAAA;AACF,IAAA;AAGM,IAAA;AAGA,IAAA;AACA,IAAA;AACJ,MAAA;AACE,QAAA;AACF,MAAA;AACD,IAAA;AAGe,IAAA;AAEE,IAAA;AACJ,EAAA;AACA,IAAA;AACP,IAAA;AACI,MAAA;AACF,MAAA;AACT,IAAA;AACF,EAAA;AACF;AAES;AACa,EAAA;AAEsB,EAAA;AAAA;AAEnC,IAAA;AACA,IAAA;AACC,IAAA;AACD,IAAA;AACC,IAAA;AACD,IAAA;AAAA;AAEA,IAAA;AACA,IAAA;AACC,IAAA;AAAA;AAED,IAAA;AACA,IAAA;AACC,IAAA;AACR,EAAA;AAEiB,EAAA;AACnB;AAEsB;AAChB,EAAA;AACW,IAAA;AACF,IAAA;AACO,MAAA;AAClB,IAAA;AAEM,IAAA;AACA,IAAA;AACA,IAAA;AACY,IAAA;AAGf,IAAA;AAKe,MAAA;AAClB,IAAA;AAGM,IAAA;AACW,IAAA;AAGX,IAAA;AAGS,IAAA;AAGE,IAAA;AACL,MAAA;AACF,MAAA;AAAA;AACK,MAAA;AACE,QAAA;AACb,QAAA;AACF,MAAA;AACgB,MAAA;AAAA;AACH,MAAA;AACG,MAAA;AACL,QAAA;AAAA;AACX,MAAA;AACD,IAAA;AAGe,IAAA;AACN,MAAA;AACH,MAAA;AACC,MAAA;AACO,MAAA;AACR,MAAA;AAAA;AACN,IAAA;AAEc,IAAA;AAGC,IAAA;AAEE,IAAA;AACJ,EAAA;AACA,IAAA;AACP,IAAA;AACI,MAAA;AACF,MAAA;AACT,IAAA;AACF,EAAA;AACF;ADxCsB;AACA;AACA;AACA;AACA","file":"/home/runner/work/eloquent-packages/eloquent-packages/packages/react/dist/chunk-W4AIK44N.js","sourcesContent":[null,"\"use server\";\nimport { BlobServiceClient } from \"@azure/storage-blob\";\nimport { PutObjectCommand, S3Client } from \"@aws-sdk/client-s3\";\n\n\nexport async function uploadFile(formData: FormData) {\n try {\n const file = formData.get(\"file\") as File;\n if (!file) {\n throw new Error(\"No file provided\");\n }\n\n const STORAGE_CONNECTION_STRING =\n process.env.AZURE_STORAGE_CONNECTION_STRING;\n const STORAGE_CONTAINER_NAME = process.env.AZURE_STORAGE_CONTAINER_NAME;\n\n if (!STORAGE_CONNECTION_STRING || !STORAGE_CONTAINER_NAME) {\n throw new Error(\n \"STORAGE_CONNECTION_STRING or STORAGE_CONTAINER_NAME not set\"\n );\n }\n\n // Generate unique filename\n const fileExtension = file.name.split(\".\").pop()?.toLowerCase();\n const fileName = `${Date.now()}-${Math.random().toString(36).substring(2)}.${fileExtension}`;\n\n // Determine content type based on file extension\n const contentType = getContentType(fileExtension);\n\n // Convert file to buffer\n const buffer = Buffer.from(await file.arrayBuffer());\n\n // Upload to Azure Blob Storage\n const blobServiceClient = BlobServiceClient.fromConnectionString(\n STORAGE_CONNECTION_STRING\n );\n const containerClient = blobServiceClient.getContainerClient(\n STORAGE_CONTAINER_NAME\n );\n\n // Create container if it doesn't exist\n await containerClient.createIfNotExists();\n\n // Upload file with content type\n const blockBlobClient = containerClient.getBlockBlobClient(fileName);\n await blockBlobClient.upload(buffer, buffer.length, {\n blobHTTPHeaders: {\n blobContentType: contentType,\n },\n });\n\n // Get the URL\n const fileUrl = blockBlobClient.url;\n\n return { success: true, fileUrl };\n } catch (error) {\n console.error(\"Upload error:\", error);\n return {\n success: false,\n error: error instanceof Error ? error.message : \"Failed to upload file\",\n };\n }\n}\n\nfunction getContentType(fileExtension: string | undefined): string {\n if (!fileExtension) return \"application/octet-stream\";\n\n const mimeTypes: Record<string, string> = {\n // Images\n png: \"image/png\",\n jpg: \"image/jpeg\",\n jpeg: \"image/jpeg\",\n gif: \"image/gif\",\n webp: \"image/webp\",\n svg: \"image/svg+xml\",\n // Documents\n pdf: \"application/pdf\",\n doc: \"application/msword\",\n docx: \"application/vnd.openxmlformats-officedocument.wordprocessingml.document\",\n // Others\n txt: \"text/plain\",\n csv: \"text/csv\",\n json: \"application/json\",\n };\n\n return mimeTypes[fileExtension] || \"application/octet-stream\";\n}\n\nexport async function uploadToS3(formData: FormData) {\n try {\n const file = formData.get(\"file\") as File;\n if (!file) {\n throw new Error(\"No file provided\");\n }\n\n const AWS_ACCESS_KEY_ID = process.env.LINODE_ACCESS_KEY;\n const AWS_SECRET_ACCESS_KEY = process.env.LINODE_SECRET_KEY;\n const S3_BUCKET_NAME = process.env.LINODE_BUCKET_NAME;\n let S3_ENDPOINT = process.env.LINODE_OBJECT_STORAGE_ENDPOINT;\n\n if (\n !AWS_ACCESS_KEY_ID ||\n !AWS_SECRET_ACCESS_KEY ||\n !S3_BUCKET_NAME ||\n !S3_ENDPOINT\n ) {\n throw new Error(\"S3 credentials or configuration not set\");\n }\n\n // Generate unique filename\n const fileExtension = file.name.split(\".\").pop()?.toLowerCase();\n const fileName = `${Date.now()}-${Math.random().toString(36).substring(2)}.${fileExtension}`;\n\n // Determine content type based on file extension\n const contentType = getContentType(fileExtension);\n\n // Convert file to buffer\n const buffer = Buffer.from(await file.arrayBuffer());\n\n // Initialize S3 client\n const s3Client = new S3Client({\n endpoint: S3_ENDPOINT,\n region: \"us-east-1\", // Linode requires a region but it's not used with custom endpoint\n credentials: {\n accessKeyId: AWS_ACCESS_KEY_ID,\n secretAccessKey: AWS_SECRET_ACCESS_KEY,\n },\n forcePathStyle: true, // Required for Linode Object Storage\n maxAttempts: 3,\n requestHandler: {\n timeout: 10000, // 10 seconds timeout\n },\n });\n\n // Upload to S3\n const command = new PutObjectCommand({\n Bucket: S3_BUCKET_NAME,\n Key: fileName,\n Body: buffer,\n ContentType: contentType,\n ACL: \"public-read\", // Make the object publicly readable\n });\n\n await s3Client.send(command);\n\n // Construct the URL using the endpoint\n const fileUrl = `${S3_ENDPOINT}/${S3_BUCKET_NAME}/${fileName}`;\n\n return { success: true, fileUrl };\n } catch (error) {\n console.error(\"Upload error:\", JSON.stringify(error, null, 2));\n return {\n success: false,\n error: error instanceof Error ? error.message : \"Failed to upload file\",\n };\n }\n}\n"]}
|
package/dist/chunk-ZCBGKBZS.js
DELETED
|
@@ -1,36 +0,0 @@
|
|
|
1
|
-
"use strict";Object.defineProperty(exports, "__esModule", {value: true});// lib/utils.ts
|
|
2
|
-
var _clsx = require('clsx');
|
|
3
|
-
var _tailwindmerge = require('tailwind-merge');
|
|
4
|
-
function cn(...inputs) {
|
|
5
|
-
return _tailwindmerge.twMerge.call(void 0, _clsx.clsx.call(void 0, inputs));
|
|
6
|
-
}
|
|
7
|
-
function getBadgeVariant(status) {
|
|
8
|
-
return status === "open" ? "default" : "secondary";
|
|
9
|
-
}
|
|
10
|
-
|
|
11
|
-
// components/form-controls/icon.tsx
|
|
12
|
-
var _lucidereact = require('lucide-react');
|
|
13
|
-
var _react = require('react');
|
|
14
|
-
var _jsxruntime = require('react/jsx-runtime');
|
|
15
|
-
var Icon = _react.memo.call(void 0, ({ name, className, strokeWidth, size }) => {
|
|
16
|
-
const IconComponent = _lucidereact.icons[name];
|
|
17
|
-
if (!IconComponent) {
|
|
18
|
-
return null;
|
|
19
|
-
}
|
|
20
|
-
return /* @__PURE__ */ _jsxruntime.jsx.call(void 0,
|
|
21
|
-
IconComponent,
|
|
22
|
-
{
|
|
23
|
-
className: cn(className),
|
|
24
|
-
strokeWidth: strokeWidth || 2.5,
|
|
25
|
-
size
|
|
26
|
-
}
|
|
27
|
-
);
|
|
28
|
-
});
|
|
29
|
-
Icon.displayName = "Icon";
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
exports.cn = cn; exports.getBadgeVariant = getBadgeVariant; exports.Icon = Icon;
|
|
36
|
-
//# sourceMappingURL=chunk-ZCBGKBZS.js.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"sources":["/home/runner/work/eloquent-packages/eloquent-packages/packages/react/dist/chunk-ZCBGKBZS.js","../lib/utils.ts","../components/form-controls/icon.tsx"],"names":[],"mappings":"AAAA;ACAA,4BAAsC;AACtC,+CAAwB;AAEjB,SAAS,EAAA,CAAA,GAAM,MAAA,EAAsB;AAC1C,EAAA,OAAO,oCAAA,wBAAQ,MAAW,CAAC,CAAA;AAC7B;AAEO,SAAS,eAAA,CAAgB,MAAA,EAAgB;AAC9C,EAAA,OAAO,OAAA,IAAW,OAAA,EAAS,UAAA,EAAY,WAAA;AACzC;ADAA;AACA;AETA,2CAAsB;AACtB,8BAAqB;AAiBjB,+CAAA;AARG,IAAM,KAAA,EAAO,yBAAA,CAAM,EAAE,IAAA,EAAM,SAAA,EAAW,WAAA,EAAa,KAAK,CAAA,EAAA,GAAiB;AAC9E,EAAA,MAAM,cAAA,EAAgB,kBAAA,CAAM,IAAI,CAAA;AAEhC,EAAA,GAAA,CAAI,CAAC,aAAA,EAAe;AAClB,IAAA,OAAO,IAAA;AAAA,EACT;AAEA,EAAA,uBACE,6BAAA;AAAA,IAAC,aAAA;AAAA,IAAA;AAAA,MACC,SAAA,EAAW,EAAA,CAAG,SAAS,CAAA;AAAA,MACvB,WAAA,EAAa,YAAA,GAAe,GAAA;AAAA,MAC5B;AAAA,IAAA;AAAA,EACF,CAAA;AAEJ,CAAC,CAAA;AAED,IAAA,CAAK,YAAA,EAAc,MAAA;AFEnB;AACA;AACE;AACA;AACA;AACF,gFAAC","file":"/home/runner/work/eloquent-packages/eloquent-packages/packages/react/dist/chunk-ZCBGKBZS.js","sourcesContent":[null,"import { type ClassValue, clsx } from \"clsx\";\nimport { twMerge } from \"tailwind-merge\";\n\nexport function cn(...inputs: ClassValue[]) {\n return twMerge(clsx(inputs));\n}\n\nexport function getBadgeVariant(status: string) {\n return status === \"open\" ? \"default\" : \"secondary\";\n}\n\n// export function getColspanByFieldType(fieldType: EntityFieldType) {\n// return fieldType === \"text\" ? \"col-span-2\" : \"\";\n// }\n","import { cn } from \"../../lib/utils\";\nimport { icons } from \"lucide-react\";\nimport { memo } from \"react\";\n\nexport type IconProps = {\n name: keyof typeof icons;\n size?: number;\n className?: string;\n strokeWidth?: number;\n};\n\nexport const Icon = memo(({ name, className, strokeWidth, size }: IconProps) => {\n const IconComponent = icons[name];\n\n if (!IconComponent) {\n return null;\n }\n\n return (\n <IconComponent\n className={cn(className)}\n strokeWidth={strokeWidth || 2.5}\n size={size}\n />\n );\n});\n\nIcon.displayName = \"Icon\";\n"]}
|