powr-sdk-api 4.5.0 → 4.5.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/config.js CHANGED
@@ -1,8 +1,26 @@
1
1
  "use strict";
2
2
 
3
+ var _process$env$GOOGLE_C;
4
+ require('dotenv').config();
3
5
  const config = {
4
6
  mongoUri: process.env.POWR_DB_URI,
5
7
  jwtToken: process.env.JWT_TOKEN,
6
- projectId: process.env.PROJECT_ID
8
+ projectId: process.env.PROJECT_ID,
9
+ storageBucket: process.env.STORAGE_BUCKET
7
10
  };
8
- module.exports = config;
11
+ const credentials = {
12
+ type: process.env.GOOGLE_CLOUD_TYPE,
13
+ project_id: process.env.GOOGLE_CLOUD_PROJECT_ID,
14
+ private_key_id: process.env.GOOGLE_CLOUD_PRIVATE_KEY_ID,
15
+ private_key: (_process$env$GOOGLE_C = process.env.GOOGLE_CLOUD_PRIVATE_KEY) === null || _process$env$GOOGLE_C === void 0 ? void 0 : _process$env$GOOGLE_C.replace(/\\n/g, "\n"),
16
+ client_email: process.env.GOOGLE_CLOUD_CLIENT_EMAIL,
17
+ client_id: process.env.GOOGLE_CLOUD_CLIENT_ID,
18
+ auth_uri: process.env.GOOGLE_CLOUD_AUTH_URI,
19
+ token_uri: process.env.GOOGLE_CLOUD_TOKEN_URI,
20
+ auth_provider_x509_cert_url: process.env.GOOGLE_CLOUD_AUTH_PROVIDER_X509_CERT_URL,
21
+ client_x509_cert_url: process.env.GOOGLE_CLOUD_CLIENT_X509_CERT_URL
22
+ };
23
+ module.exports = {
24
+ config,
25
+ credentials
26
+ };
@@ -1,12 +1,16 @@
1
1
  "use strict";
2
2
 
3
3
  const express = require('express');
4
- const router = express.Router();
5
4
  const multer = require('multer');
6
5
  // const { uploadToGCS, bucket } = require('../uploadToGCS')
7
- // const { bucket, uploadToGCS, bucketName } = require('../uploadToGCS.js')
6
+ const {
7
+ bucket,
8
+ uploadToGCS,
9
+ bucketName
10
+ } = require('../services/uploadToGCS.js');
8
11
  // const uploadToGCS = require('../uploadToGCS.js')
9
12
 
13
+ const router = express.Router();
10
14
  const upload = multer();
11
15
  router.post('/', upload.single('file'), async (req, res) => {
12
16
  try {
@@ -16,13 +20,11 @@ router.post('/', upload.single('file'), async (req, res) => {
16
20
  message: 'No file uploaded'
17
21
  });
18
22
  }
19
-
20
- // const url = await uploadToGCS(req.file);
21
-
23
+ const url = await uploadToGCS(req.file);
22
24
  return res.status(200).json({
23
25
  status: 'success',
24
- message: 'File uploaded successfully'
25
- // url,
26
+ message: 'File uploaded successfully',
27
+ url
26
28
  });
27
29
  } catch (err) {
28
30
  console.error('Error uploading file:', err);
@@ -35,13 +37,12 @@ router.post('/', upload.single('file'), async (req, res) => {
35
37
  });
36
38
  router.get('/', async (req, res) => {
37
39
  try {
38
- // const [files] = await bucket.getFiles();
39
- // const bucketName = bucket.name;
40
- // const imageUrls = files.map(file => `https://storage.googleapis.com/${bucketName}/${file.name}`);
41
-
40
+ const [files] = await bucket.getFiles();
41
+ const bucketName = bucket.name;
42
+ const imageUrls = files.map(file => `https://storage.googleapis.com/${bucketName}/${file.name}`);
42
43
  return res.status(200).json({
43
- message: 'Images fetched successfully'
44
- // images: imageUrls,
44
+ message: 'Images fetched successfully',
45
+ images: imageUrls
45
46
  });
46
47
  } catch (err) {
47
48
  console.error('Error listing files from GCS:', err);
@@ -0,0 +1,53 @@
1
+ "use strict";
2
+
3
+ const {
4
+ Storage
5
+ } = require('@google-cloud/storage');
6
+ const {
7
+ format
8
+ } = require('date-fns');
9
+ const {
10
+ config,
11
+ credentials
12
+ } = require('../config');
13
+ const storage = new Storage({
14
+ projectId: config.projectId,
15
+ credentials: credentials
16
+ });
17
+ const bucketName = config.storageBucket;
18
+ const bucket = storage.bucket(bucketName);
19
+ console.log('Bucket:', bucketName);
20
+ console.log('Bucket:', bucket);
21
+ async function uploadToGCS(file) {
22
+ const timestamp = format(new Date(), 'yyyyMMdd-HHmmss');
23
+ const cleanFileName = file.originalname.replace(/[^a-zA-Z0-9.-]/g, '-');
24
+ const filename = `${timestamp}-${cleanFileName}`;
25
+ const blob = bucket.file(filename);
26
+ try {
27
+ console.log('Preparing to upload file:', {
28
+ name: file.originalname,
29
+ type: file.mimetype
30
+ });
31
+ const buffer = file.buffer;
32
+ console.log('Uploading file to GCS...');
33
+ await blob.save(buffer, {
34
+ contentType: file.mimetype,
35
+ metadata: {
36
+ contentType: file.mimetype
37
+ }
38
+ });
39
+ await blob.makePublic();
40
+ console.log('File made public successfully');
41
+ const publicUrl = `https://storage.googleapis.com/${bucketName}/${filename}`;
42
+ console.log('File uploaded and public URL generated:', publicUrl);
43
+ return publicUrl;
44
+ } catch (error) {
45
+ console.error('Error uploading file to GCS:', error);
46
+ throw new Error(`Failed to upload file to GCS: ${error instanceof Error ? error.message : 'Unknown error'}`);
47
+ }
48
+ }
49
+ module.exports = {
50
+ uploadToGCS,
51
+ bucket,
52
+ bucketName
53
+ };
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "powr-sdk-api",
3
- "version": "4.5.0",
3
+ "version": "4.5.3",
4
4
  "description": "Shared API core library for PowrStack projects. Zero dependencies - works with Express, Next.js API routes, and other frameworks. All features are optional and install only what you need.",
5
5
  "main": "dist/index.js",
6
6
  "types": "dist/index.d.ts",
@@ -43,6 +43,7 @@
43
43
  "optionalDependencies": {
44
44
  "@aws-sdk/client-s3": "^3.787.0",
45
45
  "@google-cloud/storage": "^7.16.0",
46
+ "date-fns": "^4.1.0",
46
47
  "axios": "^1.6.0",
47
48
  "bcrypt": "^5.1.1",
48
49
  "cron-parser": "^4.9.0",
@@ -1,78 +0,0 @@
1
- "use strict";
2
-
3
- const winston = require('winston');
4
- const {
5
- Storage
6
- } = require('@google-cloud/storage');
7
-
8
- // Create GCS client using environment variables for credentials
9
- const storage = new Storage({
10
- projectId: process.env.GCS_PROJECT_ID,
11
- credentials: {
12
- client_email: process.env.GCS_CLIENT_EMAIL,
13
- private_key: process.env.GCS_PRIVATE_KEY ? process.env.GCS_PRIVATE_KEY.replace(/\\n/g, '\n') : undefined
14
- }
15
- });
16
-
17
- /**
18
- * Custom Winston transport for logging to Google Cloud Storage
19
- */
20
- class GCSTransport extends winston.Transport {
21
- constructor(opts) {
22
- super(opts);
23
- this.bucket = opts.bucket;
24
- this.prefix = opts.prefix || '';
25
- this.buffer = [];
26
- this.bufferSize = opts.bufferSize || 100;
27
- this.flushInterval = opts.flushInterval || 5000;
28
- this.setupFlushInterval();
29
- }
30
- setupFlushInterval() {
31
- setInterval(() => {
32
- this.flush();
33
- }, this.flushInterval);
34
- }
35
- async flush() {
36
- if (this.buffer.length === 0) return;
37
- const logs = this.buffer.splice(0, this.buffer.length);
38
- const date = new Date().toISOString().split('T')[0];
39
- const filename = `${this.prefix}/${date}/${Date.now()}.json`;
40
- try {
41
- const bucket = storage.bucket(this.bucket);
42
- const file = bucket.file(filename);
43
- await file.save(JSON.stringify(logs), {
44
- contentType: 'application/json',
45
- metadata: {
46
- contentType: 'application/json'
47
- }
48
- });
49
- } catch (error) {
50
- console.error('Failed to write logs to Google Cloud Storage:', {
51
- error: error.message,
52
- code: error.code,
53
- bucket: this.bucket,
54
- filename: filename,
55
- projectId: process.env.GCS_PROJECT_ID,
56
- stack: error.stack
57
- });
58
- // Put the logs back in the buffer
59
- this.buffer.unshift(...logs);
60
- }
61
- }
62
- log(info, callback) {
63
- setImmediate(() => {
64
- this.emit('logged', info);
65
- });
66
- this.buffer.push({
67
- timestamp: new Date().toISOString(),
68
- ...info
69
- });
70
- if (this.buffer.length >= this.bufferSize) {
71
- this.flush();
72
- }
73
- callback();
74
- }
75
- }
76
- module.exports = {
77
- GCSTransport
78
- };
package/dist/logger/s3.js DELETED
@@ -1,78 +0,0 @@
1
- "use strict";
2
-
3
- const winston = require('winston');
4
- const {
5
- S3Client,
6
- PutObjectCommand
7
- } = require('@aws-sdk/client-s3');
8
-
9
- // Create S3 client
10
- const s3Client = new S3Client({
11
- region: process.env.AWS_REGION,
12
- credentials: {
13
- accessKeyId: process.env.AWS_ACCESS_KEY_ID,
14
- secretAccessKey: process.env.AWS_SECRET_ACCESS_KEY
15
- }
16
- });
17
-
18
- /**
19
- * Custom Winston transport for logging to S3
20
- */
21
- class S3Transport extends winston.Transport {
22
- constructor(opts) {
23
- super(opts);
24
- this.bucket = opts.bucket;
25
- this.prefix = opts.prefix || '';
26
- this.buffer = [];
27
- this.bufferSize = opts.bufferSize || 100;
28
- this.flushInterval = opts.flushInterval || 5000;
29
- this.setupFlushInterval();
30
- }
31
- setupFlushInterval() {
32
- setInterval(() => {
33
- this.flush();
34
- }, this.flushInterval);
35
- }
36
- async flush() {
37
- if (this.buffer.length === 0) return;
38
- const logs = this.buffer.splice(0, this.buffer.length);
39
- const date = new Date().toISOString().split('T')[0];
40
- const key = `${this.prefix}/${date}/${Date.now()}.json`;
41
- try {
42
- const command = new PutObjectCommand({
43
- Bucket: this.bucket,
44
- Key: key,
45
- Body: JSON.stringify(logs),
46
- ContentType: 'application/json'
47
- });
48
- await s3Client.send(command);
49
- } catch (error) {
50
- console.error('Failed to write logs to S3:', {
51
- error: error.message,
52
- code: error.code,
53
- bucket: this.bucket,
54
- key: key,
55
- region: process.env.AWS_REGION,
56
- stack: error.stack
57
- });
58
- // Put the logs back in the buffer
59
- this.buffer.unshift(...logs);
60
- }
61
- }
62
- log(info, callback) {
63
- setImmediate(() => {
64
- this.emit('logged', info);
65
- });
66
- this.buffer.push({
67
- timestamp: new Date().toISOString(),
68
- ...info
69
- });
70
- if (this.buffer.length >= this.bufferSize) {
71
- this.flush();
72
- }
73
- callback();
74
- }
75
- }
76
- module.exports = {
77
- S3Transport
78
- };
@@ -1,76 +0,0 @@
1
- "use strict";
2
-
3
- const jwt = require("jsonwebtoken");
4
- const generateToken = user => {
5
- return jwt.sign({
6
- userId: user.userId
7
- }, process.env.JWT_SECRET, {
8
- expiresIn: process.env.JWT_EXPIRES_IN || '24h'
9
- });
10
- };
11
- const validateToken = async (req, res, next) => {
12
- try {
13
- let token;
14
-
15
- // Check if token exists in headers
16
- if (req.headers.authorization && req.headers.authorization.startsWith("Bearer")) {
17
- token = req.headers.authorization.split(" ")[1];
18
- }
19
- if (!token) {
20
- return res.error("Not authorized to access this route", 401);
21
- }
22
- try {
23
- // Verify token
24
- const decoded = jwt.verify(token, process.env.JWT_SECRET);
25
-
26
- // // Get user from token
27
- // const user = await User.findByPk(decoded.userId);
28
- // if (!user) {
29
- // return res.error('User not found', 401);
30
- // }
31
-
32
- // // Add user and userId to request object
33
- // req.user = user;
34
- req.userId = decoded.userId;
35
- next();
36
- } catch (error) {
37
- return res.error("Not authorized to access this route", 401, error);
38
- }
39
- } catch (error) {
40
- return res.error("Error authenticating user", 500, error);
41
- }
42
- };
43
- const validateAuth = (options = {}) => {
44
- const {
45
- publicPaths = ["/auth", "/", "/status", "/swagger"],
46
- publicMethods = ["OPTIONS"]
47
- } = options;
48
- return async (req, res, next) => {
49
- // Skip auth validation for public paths
50
- if (publicPaths.some(path => req.path.startsWith(path))) {
51
- return next();
52
- }
53
-
54
- // Skip auth validation for public methods
55
- if (publicMethods.includes(req.method)) {
56
- return next();
57
- }
58
-
59
- // Use the original validateToken middleware for protected routes
60
- return validateToken(req, res, next);
61
- };
62
- };
63
-
64
- // exports.authorize = (...roles) => {
65
- // return (req, res, next) => {
66
- // if (!roles.includes(req.user.role)) {
67
- // return res.error(`User role ${req.user.role} is not authorized to access this route`, 403);
68
- // }
69
- // next();
70
- // };
71
- // };
72
-
73
- module.exports = {
74
- validateAuth,
75
- generateToken
76
- };
@@ -1,45 +0,0 @@
1
- "use strict";
2
-
3
- /**
4
- * Logging middleware for Express applications
5
- */
6
-
7
- const logger = (req, res, next) => {
8
- const startTime = Date.now();
9
-
10
- // Log request with redacted sensitive information
11
- try {
12
- const logBody = {
13
- ...req.body
14
- };
15
- if (logBody.password) logBody.password = "[REDACTED]";
16
- if (logBody.token) logBody.token = "[REDACTED]";
17
- console.log('Request:', {
18
- timestamp: new Date().toISOString(),
19
- method: req.method,
20
- url: req.originalUrl,
21
- ip: req.ip,
22
- userAgent: req.get('user-agent'),
23
- body: logBody,
24
- query: req.query
25
- });
26
- } catch (error) {
27
- console.error("Error logging request:", error);
28
- }
29
-
30
- // Log response with timing information
31
- res.on('finish', () => {
32
- const responseTime = Date.now() - startTime;
33
- console.log('Response:', {
34
- timestamp: new Date().toISOString(),
35
- method: req.method,
36
- url: req.originalUrl,
37
- statusCode: res.statusCode,
38
- responseTime: `${responseTime}ms`,
39
- userAgent: req.get('user-agent'),
40
- ip: req.ip
41
- });
42
- });
43
- next();
44
- };
45
- module.exports = logger;
@@ -1,53 +0,0 @@
1
- "use strict";
2
-
3
- const logger = require('../services/logger');
4
- const responseFormatter = (req, res, next) => {
5
- // Add success method to response object
6
- res.success = function (message = 'Success', statusCode = 200, data = null) {
7
- const response = {
8
- success: true,
9
- message
10
- };
11
-
12
- // Only include data if it's not null
13
- if (data !== null) {
14
- response.data = data;
15
- }
16
- return this.status(statusCode).json(response);
17
- };
18
-
19
- // Add error method to response object
20
- res.error = function (message = 'Error', statusCode = 500, error = null) {
21
- const response = {
22
- success: false,
23
- message
24
- };
25
-
26
- // Only include details if they're not null
27
- if (error !== null) {
28
- response.error = error;
29
- }
30
- return this.status(statusCode).json(response);
31
- };
32
- next();
33
- };
34
- const responseLogger = (req, res, next) => {
35
- const startTime = Date.now();
36
- res.on('finish', () => {
37
- const responseTime = Date.now() - startTime;
38
- logger.info('API Response', {
39
- method: req.method,
40
- url: req.originalUrl,
41
- statusCode: res.statusCode,
42
- responseTime: `${responseTime}ms`,
43
- userAgent: req.get('user-agent'),
44
- ip: req.ip,
45
- response: res.locals.responseBody || '[not captured]'
46
- });
47
- });
48
- next();
49
- };
50
- module.exports = {
51
- responseFormatter,
52
- responseLogger
53
- };
@@ -1,35 +0,0 @@
1
- "use strict";
2
-
3
- const winston = require('winston');
4
- const {
5
- S3Transport
6
- } = require('../utils/s3');
7
-
8
- // Custom format for logs
9
- const logFormat = winston.format.combine(winston.format.timestamp(), winston.format.errors({
10
- stack: true
11
- }), winston.format.json());
12
-
13
- // Create base transports array with console transport
14
- const transports = [new winston.transports.Console()];
15
-
16
- // Add S3 transport only in production
17
- if (process.env.NODE_ENV === 'production') {
18
- transports.push(new S3Transport({
19
- bucket: process.env.AWS_LOG_BUCKET_NAME,
20
- prefix: 'logs',
21
- bufferSize: 100,
22
- flushInterval: 5000
23
- }));
24
- }
25
-
26
- // Create the logger
27
- const logger = winston.createLogger({
28
- level: process.env.LOG_LEVEL || 'info',
29
- format: logFormat,
30
- defaultMeta: {
31
- service: process.env.LOG_SERVICE_NAME
32
- },
33
- transports
34
- });
35
- module.exports = logger;
@@ -1,19 +0,0 @@
1
- "use strict";
2
-
3
- const jwt = require('jsonwebtoken');
4
-
5
- /**
6
- * Generates a JWT token for a user
7
- * @param {Object} user - The user object
8
- * @returns {string} - The generated JWT token
9
- */
10
- const generateToken = user => {
11
- return jwt.sign({
12
- userId: user.userId
13
- }, process.env.JWT_SECRET, {
14
- expiresIn: process.env.JWT_EXPIRES_IN || '24h'
15
- });
16
- };
17
- module.exports = {
18
- generateToken
19
- };
@@ -1,57 +0,0 @@
1
- "use strict";
2
-
3
- /**
4
- * Logger utility for Express applications
5
- */
6
-
7
- const winston = require('winston');
8
- const S3Transport = require('./s3-transport');
9
-
10
- /**
11
- * Creates a Winston logger instance with configured transports
12
- * @param {Object} options - Logger configuration options
13
- * @param {string} options.service - Service name for logging
14
- * @param {string} options.level - Log level (default: 'info')
15
- * @param {Object} options.s3 - S3 transport configuration
16
- * @param {string} options.s3.bucket - S3 bucket name
17
- * @param {string} options.s3.prefix - S3 key prefix
18
- * @param {number} options.s3.bufferSize - Buffer size for S3 uploads
19
- * @param {number} options.s3.flushInterval - Flush interval in milliseconds
20
- * @returns {winston.Logger} Configured Winston logger instance
21
- */
22
- const createLogger = (options = {}) => {
23
- const {
24
- service = 'api',
25
- level = process.env.LOG_LEVEL || 'info',
26
- s3 = null
27
- } = options;
28
-
29
- // Custom format for logs
30
- const logFormat = winston.format.combine(winston.format.timestamp(), winston.format.errors({
31
- stack: true
32
- }), winston.format.json());
33
-
34
- // Create base transports array with console transport
35
- const transports = [new winston.transports.Console()];
36
-
37
- // Add S3 transport if configured
38
- if (s3 && process.env.NODE_ENV === 'production') {
39
- transports.push(new S3Transport({
40
- bucket: s3.bucket,
41
- prefix: s3.prefix || 'logs',
42
- bufferSize: s3.bufferSize || 100,
43
- flushInterval: s3.flushInterval || 5000
44
- }));
45
- }
46
-
47
- // Create and return the logger
48
- return winston.createLogger({
49
- level,
50
- format: logFormat,
51
- defaultMeta: {
52
- service
53
- },
54
- transports
55
- });
56
- };
57
- module.exports = createLogger;
@@ -1,61 +0,0 @@
1
- "use strict";
2
-
3
- /**
4
- * S3 Transport for Winston logger
5
- */
6
-
7
- const winston = require('winston');
8
- const {
9
- S3Client,
10
- PutObjectCommand
11
- } = require('@aws-sdk/client-s3');
12
-
13
- /**
14
- * Custom Winston transport for logging to S3
15
- */
16
- class S3Transport extends winston.Transport {
17
- constructor(opts) {
18
- super(opts);
19
- this.bucket = opts.bucket;
20
- this.prefix = opts.prefix || '';
21
- this.buffer = [];
22
- this.bufferSize = opts.bufferSize || 100;
23
- this.flushInterval = opts.flushInterval || 5000;
24
- this.setupFlushInterval();
25
- }
26
- setupFlushInterval() {
27
- setInterval(() => {
28
- this.flush();
29
- }, this.flushInterval);
30
- }
31
- async flush() {
32
- if (this.buffer.length === 0) return;
33
- const logs = this.buffer.splice(0, this.buffer.length);
34
- const date = new Date().toISOString().split('T')[0];
35
- const key = `${this.prefix}/${date}/${Date.now()}.json`;
36
- try {
37
- const command = new PutObjectCommand({
38
- Bucket: this.bucket,
39
- Key: key,
40
- Body: JSON.stringify(logs),
41
- ContentType: 'application/json'
42
- });
43
- await this.s3Client.send(command);
44
- } catch (error) {
45
- console.error('Error flushing logs to S3:', error);
46
- // Put logs back in buffer
47
- this.buffer.unshift(...logs);
48
- }
49
- }
50
- log(info, callback) {
51
- setImmediate(() => {
52
- this.emit('logged', info);
53
- });
54
- this.buffer.push(info);
55
- if (this.buffer.length >= this.bufferSize) {
56
- this.flush();
57
- }
58
- callback();
59
- }
60
- }
61
- module.exports = S3Transport;
package/dist/utils/s3.js DELETED
@@ -1,78 +0,0 @@
1
- "use strict";
2
-
3
- const winston = require('winston');
4
- const {
5
- S3Client,
6
- PutObjectCommand
7
- } = require('@aws-sdk/client-s3');
8
-
9
- // Create S3 client
10
- const s3Client = new S3Client({
11
- region: process.env.AWS_REGION,
12
- credentials: {
13
- accessKeyId: process.env.AWS_ACCESS_KEY_ID,
14
- secretAccessKey: process.env.AWS_SECRET_ACCESS_KEY
15
- }
16
- });
17
-
18
- /**
19
- * Custom Winston transport for logging to S3
20
- */
21
- class S3Transport extends winston.Transport {
22
- constructor(opts) {
23
- super(opts);
24
- this.bucket = opts.bucket;
25
- this.prefix = opts.prefix || '';
26
- this.buffer = [];
27
- this.bufferSize = opts.bufferSize || 100;
28
- this.flushInterval = opts.flushInterval || 5000;
29
- this.setupFlushInterval();
30
- }
31
- setupFlushInterval() {
32
- setInterval(() => {
33
- this.flush();
34
- }, this.flushInterval);
35
- }
36
- async flush() {
37
- if (this.buffer.length === 0) return;
38
- const logs = this.buffer.splice(0, this.buffer.length);
39
- const date = new Date().toISOString().split('T')[0];
40
- const key = `${this.prefix}/${date}/${Date.now()}.json`;
41
- try {
42
- const command = new PutObjectCommand({
43
- Bucket: this.bucket,
44
- Key: key,
45
- Body: JSON.stringify(logs),
46
- ContentType: 'application/json'
47
- });
48
- await s3Client.send(command);
49
- } catch (error) {
50
- console.error('Failed to write logs to S3:', {
51
- error: error.message,
52
- code: error.code,
53
- bucket: this.bucket,
54
- key: key,
55
- region: process.env.AWS_REGION,
56
- stack: error.stack
57
- });
58
- // Put the logs back in the buffer
59
- this.buffer.unshift(...logs);
60
- }
61
- }
62
- log(info, callback) {
63
- setImmediate(() => {
64
- this.emit('logged', info);
65
- });
66
- this.buffer.push({
67
- timestamp: new Date().toISOString(),
68
- ...info
69
- });
70
- if (this.buffer.length >= this.bufferSize) {
71
- this.flush();
72
- }
73
- callback();
74
- }
75
- }
76
- module.exports = {
77
- S3Transport
78
- };