@remotion/lambda 3.3.6 → 3.3.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/admin/make-layer-public.js +23 -1
- package/dist/api/deploy-site.d.ts +5 -0
- package/dist/api/deploy-site.js +39 -16
- package/dist/api/upload-dir.d.ts +4 -3
- package/dist/api/upload-dir.js +33 -19
- package/dist/cli/commands/sites/create.js +12 -2
- package/dist/cli/helpers/progress-bar.d.ts +8 -1
- package/dist/cli/helpers/progress-bar.js +17 -1
- package/dist/cli/index.js +6 -1
- package/dist/functions/helpers/get-chromium-executable-path.js +2 -1
- package/dist/functions/helpers/get-current-region.d.ts +1 -1
- package/dist/pricing/aws-regions.d.ts +1 -1
- package/dist/pricing/aws-regions.js +11 -0
- package/dist/pricing/price-per-1-s.js +413 -39
- package/dist/shared/get-etag.d.ts +1 -0
- package/dist/shared/get-etag.js +24 -0
- package/dist/shared/get-s3-operations.d.ts +10 -0
- package/dist/shared/get-s3-operations.js +36 -0
- package/dist/shared/hosted-layers.js +368 -60
- package/dist/shared/read-dir.d.ts +9 -0
- package/dist/shared/read-dir.js +57 -0
- package/package.json +6 -6
- package/remotionlambda.zip +0 -0
|
@@ -20,6 +20,17 @@ const layerInfo = {
|
|
|
20
20
|
'us-east-1': [],
|
|
21
21
|
'us-east-2': [],
|
|
22
22
|
'us-west-2': [],
|
|
23
|
+
'af-south-1': [],
|
|
24
|
+
'ap-east-1': [],
|
|
25
|
+
'ap-northeast-2': [],
|
|
26
|
+
'ap-northeast-3': [],
|
|
27
|
+
'ca-central-1': [],
|
|
28
|
+
'eu-north-1': [],
|
|
29
|
+
'eu-south-1': [],
|
|
30
|
+
'eu-west-3': [],
|
|
31
|
+
'me-south-1': [],
|
|
32
|
+
'sa-east-1': [],
|
|
33
|
+
'us-west-1': [],
|
|
23
34
|
},
|
|
24
35
|
x86_64: {
|
|
25
36
|
'ap-northeast-1': [],
|
|
@@ -32,6 +43,17 @@ const layerInfo = {
|
|
|
32
43
|
'us-east-1': [],
|
|
33
44
|
'us-east-2': [],
|
|
34
45
|
'us-west-2': [],
|
|
46
|
+
'af-south-1': [],
|
|
47
|
+
'ap-east-1': [],
|
|
48
|
+
'ap-northeast-2': [],
|
|
49
|
+
'ap-northeast-3': [],
|
|
50
|
+
'ca-central-1': [],
|
|
51
|
+
'eu-north-1': [],
|
|
52
|
+
'eu-south-1': [],
|
|
53
|
+
'eu-west-3': [],
|
|
54
|
+
'me-south-1': [],
|
|
55
|
+
'sa-east-1': [],
|
|
56
|
+
'us-west-1': [],
|
|
35
57
|
},
|
|
36
58
|
};
|
|
37
59
|
const makeLayerPublic = async () => {
|
|
@@ -43,7 +65,7 @@ const makeLayerPublic = async () => {
|
|
|
43
65
|
const { Version, LayerArn } = await (0, aws_clients_1.getLambdaClient)(region).send(new client_lambda_1.PublishLayerVersionCommand({
|
|
44
66
|
Content: {
|
|
45
67
|
S3Bucket: 'remotionlambda-binaries-' + region,
|
|
46
|
-
S3Key: `remotion-layer-${layer}-
|
|
68
|
+
S3Key: `remotion-layer-${layer}-v9-${architecture}.zip`,
|
|
47
69
|
},
|
|
48
70
|
LayerName: layerName,
|
|
49
71
|
LicenseInfo: layer === 'chromium'
|
|
@@ -18,6 +18,11 @@ export declare type DeploySiteInput = {
|
|
|
18
18
|
export declare type DeploySiteOutput = Promise<{
|
|
19
19
|
serveUrl: string;
|
|
20
20
|
siteName: string;
|
|
21
|
+
stats: {
|
|
22
|
+
uploadedFiles: number;
|
|
23
|
+
deletedFiles: number;
|
|
24
|
+
untouchedFiles: number;
|
|
25
|
+
};
|
|
21
26
|
}>;
|
|
22
27
|
/**
|
|
23
28
|
* @description Deploys a Remotion project to an S3 bucket to prepare it for rendering on AWS Lambda.
|
package/dist/api/deploy-site.js
CHANGED
|
@@ -1,10 +1,11 @@
|
|
|
1
1
|
"use strict";
|
|
2
2
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
3
|
exports.deploySite = void 0;
|
|
4
|
-
const
|
|
4
|
+
const io_1 = require("../functions/helpers/io");
|
|
5
5
|
const bundle_site_1 = require("../shared/bundle-site");
|
|
6
6
|
const constants_1 = require("../shared/constants");
|
|
7
7
|
const get_account_id_1 = require("../shared/get-account-id");
|
|
8
|
+
const get_s3_operations_1 = require("../shared/get-s3-operations");
|
|
8
9
|
const make_s3_url_1 = require("../shared/make-s3-url");
|
|
9
10
|
const random_hash_1 = require("../shared/random-hash");
|
|
10
11
|
const validate_aws_region_1 = require("../shared/validate-aws-region");
|
|
@@ -28,37 +29,54 @@ const deploySite = async ({ bucketName, entryPoint, siteName, options, region, }
|
|
|
28
29
|
(0, validate_bucketname_1.validateBucketName)(bucketName, { mustStartWithRemotion: true });
|
|
29
30
|
const siteId = siteName !== null && siteName !== void 0 ? siteName : (0, random_hash_1.randomHash)();
|
|
30
31
|
(0, validate_site_name_1.validateSiteName)(siteId);
|
|
32
|
+
const accountId = await (0, get_account_id_1.getAccountId)({ region });
|
|
31
33
|
const bucketExists = await (0, bucket_exists_1.bucketExistsInRegion)({
|
|
32
34
|
bucketName,
|
|
33
35
|
region,
|
|
34
|
-
expectedBucketOwner:
|
|
36
|
+
expectedBucketOwner: accountId,
|
|
35
37
|
});
|
|
36
38
|
if (!bucketExists) {
|
|
37
39
|
throw new Error(`No bucket with the name ${bucketName} exists`);
|
|
38
40
|
}
|
|
39
41
|
const subFolder = (0, constants_1.getSitesKey)(siteId);
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
|
|
47
|
-
|
|
48
|
-
|
|
49
|
-
|
|
50
|
-
|
|
51
|
-
|
|
42
|
+
const [files, bundled] = await Promise.all([
|
|
43
|
+
(0, io_1.lambdaLs)({
|
|
44
|
+
bucketName,
|
|
45
|
+
expectedBucketOwner: accountId,
|
|
46
|
+
region,
|
|
47
|
+
prefix: subFolder,
|
|
48
|
+
}),
|
|
49
|
+
(0, bundle_site_1.bundleSite)(entryPoint, (_a = options === null || options === void 0 ? void 0 : options.onBundleProgress) !== null && _a !== void 0 ? _a : (() => undefined), {
|
|
50
|
+
publicPath: `/${subFolder}/`,
|
|
51
|
+
webpackOverride: (_b = options === null || options === void 0 ? void 0 : options.webpackOverride) !== null && _b !== void 0 ? _b : ((f) => f),
|
|
52
|
+
enableCaching: (_c = options === null || options === void 0 ? void 0 : options.enableCaching) !== null && _c !== void 0 ? _c : true,
|
|
53
|
+
publicDir: options === null || options === void 0 ? void 0 : options.publicDir,
|
|
54
|
+
rootDir: options === null || options === void 0 ? void 0 : options.rootDir,
|
|
55
|
+
}),
|
|
56
|
+
]);
|
|
57
|
+
const { toDelete, toUpload, existingCount } = await (0, get_s3_operations_1.getS3DiffOperations)({
|
|
58
|
+
objects: files,
|
|
59
|
+
bundle: bundled,
|
|
60
|
+
prefix: subFolder,
|
|
52
61
|
});
|
|
53
62
|
await Promise.all([
|
|
54
63
|
(0, upload_dir_1.uploadDir)({
|
|
55
64
|
bucket: bucketName,
|
|
56
65
|
region,
|
|
57
|
-
|
|
66
|
+
localDir: bundled,
|
|
58
67
|
onProgress: (_d = options === null || options === void 0 ? void 0 : options.onUploadProgress) !== null && _d !== void 0 ? _d : (() => undefined),
|
|
59
|
-
|
|
68
|
+
keyPrefix: subFolder,
|
|
60
69
|
privacy: 'public',
|
|
70
|
+
toUpload,
|
|
61
71
|
}),
|
|
72
|
+
Promise.all(toDelete.map((d) => {
|
|
73
|
+
return (0, io_1.lambdaDeleteFile)({
|
|
74
|
+
bucketName,
|
|
75
|
+
customCredentials: null,
|
|
76
|
+
key: d.Key,
|
|
77
|
+
region,
|
|
78
|
+
});
|
|
79
|
+
})),
|
|
62
80
|
(0, enable_s3_website_1.enableS3Website)({
|
|
63
81
|
region,
|
|
64
82
|
bucketName,
|
|
@@ -67,6 +85,11 @@ const deploySite = async ({ bucketName, entryPoint, siteName, options, region, }
|
|
|
67
85
|
return {
|
|
68
86
|
serveUrl: (0, make_s3_url_1.makeS3ServeUrl)({ bucketName, subFolder, region }),
|
|
69
87
|
siteName: siteId,
|
|
88
|
+
stats: {
|
|
89
|
+
uploadedFiles: toUpload.length,
|
|
90
|
+
deletedFiles: toDelete.length,
|
|
91
|
+
untouchedFiles: existingCount,
|
|
92
|
+
},
|
|
70
93
|
};
|
|
71
94
|
};
|
|
72
95
|
exports.deploySite = deploySite;
|
package/dist/api/upload-dir.d.ts
CHANGED
|
@@ -11,11 +11,12 @@ export declare type MockFile = {
|
|
|
11
11
|
content: string;
|
|
12
12
|
};
|
|
13
13
|
export declare const getDirFiles: (entry: string) => MockFile[];
|
|
14
|
-
export declare const uploadDir: ({ bucket, region,
|
|
14
|
+
export declare const uploadDir: ({ bucket, region, localDir, onProgress, keyPrefix, privacy, toUpload, }: {
|
|
15
15
|
bucket: string;
|
|
16
16
|
region: AwsRegion;
|
|
17
|
-
|
|
18
|
-
|
|
17
|
+
localDir: string;
|
|
18
|
+
keyPrefix: string;
|
|
19
19
|
onProgress: (progress: UploadDirProgress) => void;
|
|
20
20
|
privacy: Privacy;
|
|
21
|
+
toUpload: string[];
|
|
21
22
|
}) => Promise<void>;
|
package/dist/api/upload-dir.js
CHANGED
|
@@ -15,31 +15,45 @@ const getDirFiles = (entry) => {
|
|
|
15
15
|
throw new TypeError('should only be executed in test ' + JSON.stringify(entry));
|
|
16
16
|
};
|
|
17
17
|
exports.getDirFiles = getDirFiles;
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
|
|
23
|
-
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
}
|
|
33
|
-
return
|
|
34
|
-
}
|
|
35
|
-
|
|
18
|
+
async function getFiles(directory, originalDirectory, toUpload) {
|
|
19
|
+
const dirents = await fs_1.promises.readdir(directory, { withFileTypes: true });
|
|
20
|
+
const _files = await Promise.all(dirents
|
|
21
|
+
.map((dirent) => {
|
|
22
|
+
const res = path_1.default.resolve(directory, dirent.name);
|
|
23
|
+
return [dirent, res];
|
|
24
|
+
})
|
|
25
|
+
.filter(([dirent, res]) => {
|
|
26
|
+
const relative = path_1.default.relative(originalDirectory, res);
|
|
27
|
+
if (dirent.isDirectory()) {
|
|
28
|
+
return true;
|
|
29
|
+
}
|
|
30
|
+
if (!toUpload.includes(relative)) {
|
|
31
|
+
return false;
|
|
32
|
+
}
|
|
33
|
+
return true;
|
|
34
|
+
})
|
|
35
|
+
.map(async ([dirent, res]) => {
|
|
36
|
+
const { size } = await fs_1.promises.stat(res);
|
|
37
|
+
return dirent.isDirectory()
|
|
38
|
+
? getFiles(res, originalDirectory, toUpload)
|
|
39
|
+
: [
|
|
40
|
+
{
|
|
41
|
+
name: res,
|
|
42
|
+
size,
|
|
43
|
+
},
|
|
44
|
+
];
|
|
45
|
+
}));
|
|
46
|
+
return _files.flat(1);
|
|
47
|
+
}
|
|
48
|
+
const uploadDir = async ({ bucket, region, localDir, onProgress, keyPrefix, privacy, toUpload, }) => {
|
|
49
|
+
const files = await getFiles(localDir, localDir, toUpload);
|
|
36
50
|
const progresses = {};
|
|
37
51
|
for (const file of files) {
|
|
38
52
|
progresses[file.name] = 0;
|
|
39
53
|
}
|
|
40
54
|
const client = (0, aws_clients_1.getS3Client)(region, null);
|
|
41
55
|
const uploads = files.map(async (filePath) => {
|
|
42
|
-
const Key = (0, make_s3_key_1.makeS3Key)(
|
|
56
|
+
const Key = (0, make_s3_key_1.makeS3Key)(keyPrefix, localDir, filePath.name);
|
|
43
57
|
const Body = (0, fs_1.createReadStream)(filePath.name);
|
|
44
58
|
const ContentType = mime_types_1.default.lookup(Key) || 'application/octet-stream';
|
|
45
59
|
const ACL = privacy === 'no-acl'
|
|
@@ -7,6 +7,7 @@ exports.sitesCreateSubcommand = exports.SITES_CREATE_SUBCOMMAND = void 0;
|
|
|
7
7
|
const cli_1 = require("@remotion/cli");
|
|
8
8
|
const fs_1 = require("fs");
|
|
9
9
|
const path_1 = __importDefault(require("path"));
|
|
10
|
+
const remotion_1 = require("remotion");
|
|
10
11
|
const deploy_site_1 = require("../../../api/deploy-site");
|
|
11
12
|
const get_or_create_bucket_1 = require("../../../api/get-or-create-bucket");
|
|
12
13
|
const constants_1 = require("../../../shared/constants");
|
|
@@ -57,6 +58,7 @@ const sitesCreateSubcommand = async (args, remotionRoot) => {
|
|
|
57
58
|
doneIn: null,
|
|
58
59
|
totalSize: null,
|
|
59
60
|
sizeUploaded: 0,
|
|
61
|
+
stats: null,
|
|
60
62
|
},
|
|
61
63
|
};
|
|
62
64
|
const updateProgress = () => {
|
|
@@ -79,7 +81,7 @@ const sitesCreateSubcommand = async (args, remotionRoot) => {
|
|
|
79
81
|
updateProgress();
|
|
80
82
|
const bundleStart = Date.now();
|
|
81
83
|
const uploadStart = Date.now();
|
|
82
|
-
const { serveUrl, siteName } = await (0, deploy_site_1.deploySite)({
|
|
84
|
+
const { serveUrl, siteName, stats } = await (0, deploy_site_1.deploySite)({
|
|
83
85
|
entryPoint: absoluteFile,
|
|
84
86
|
siteName: desiredSiteName,
|
|
85
87
|
bucketName,
|
|
@@ -95,6 +97,7 @@ const sitesCreateSubcommand = async (args, remotionRoot) => {
|
|
|
95
97
|
sizeUploaded: p.sizeUploaded,
|
|
96
98
|
totalSize: p.totalSize,
|
|
97
99
|
doneIn: null,
|
|
100
|
+
stats: null,
|
|
98
101
|
};
|
|
99
102
|
updateProgress();
|
|
100
103
|
},
|
|
@@ -108,6 +111,11 @@ const sitesCreateSubcommand = async (args, remotionRoot) => {
|
|
|
108
111
|
sizeUploaded: 1,
|
|
109
112
|
totalSize: 1,
|
|
110
113
|
doneIn: uploadDuration,
|
|
114
|
+
stats: {
|
|
115
|
+
addedFiles: stats.uploadedFiles,
|
|
116
|
+
removedFiles: stats.deletedFiles,
|
|
117
|
+
untouchedFiles: stats.untouchedFiles,
|
|
118
|
+
},
|
|
111
119
|
};
|
|
112
120
|
updateProgress();
|
|
113
121
|
log_1.Log.info();
|
|
@@ -117,6 +125,8 @@ const sitesCreateSubcommand = async (args, remotionRoot) => {
|
|
|
117
125
|
log_1.Log.info(`Site Name: ${siteName}`);
|
|
118
126
|
log_1.Log.info();
|
|
119
127
|
log_1.Log.info(cli_1.CliInternals.chalk.blueBright('ℹ️ If you make changes to your code, you need to redeploy the site. You can overwrite the existing site by running:'));
|
|
120
|
-
log_1.Log.info(cli_1.CliInternals.chalk.blueBright(
|
|
128
|
+
log_1.Log.info(cli_1.CliInternals.chalk.blueBright(['npx remotion lambda sites create', args[0], `--site-name=${siteName}`]
|
|
129
|
+
.filter(remotion_1.Internals.truthy)
|
|
130
|
+
.join(' ')));
|
|
121
131
|
};
|
|
122
132
|
exports.sitesCreateSubcommand = sitesCreateSubcommand;
|
|
@@ -9,9 +9,16 @@ export declare type BucketCreationProgress = {
|
|
|
9
9
|
doneIn: number | null;
|
|
10
10
|
};
|
|
11
11
|
export declare const makeBucketProgress: ({ bucketCreated, websiteEnabled, doneIn, }: BucketCreationProgress) => string;
|
|
12
|
+
declare type UploadStats = {
|
|
13
|
+
addedFiles: number;
|
|
14
|
+
removedFiles: number;
|
|
15
|
+
untouchedFiles: number;
|
|
16
|
+
};
|
|
12
17
|
export declare type DeployToS3Progress = {
|
|
13
18
|
sizeUploaded: number;
|
|
14
19
|
totalSize: number | null;
|
|
15
20
|
doneIn: number | null;
|
|
21
|
+
stats: UploadStats | null;
|
|
16
22
|
};
|
|
17
|
-
export declare const makeDeployProgressBar: ({ sizeUploaded, totalSize, doneIn, }: DeployToS3Progress) => string;
|
|
23
|
+
export declare const makeDeployProgressBar: ({ sizeUploaded, totalSize, doneIn, stats, }: DeployToS3Progress) => string;
|
|
24
|
+
export {};
|
|
@@ -30,7 +30,22 @@ const makeBucketProgress = ({ bucketCreated, websiteEnabled, doneIn, }) => {
|
|
|
30
30
|
].join(' ');
|
|
31
31
|
};
|
|
32
32
|
exports.makeBucketProgress = makeBucketProgress;
|
|
33
|
-
const
|
|
33
|
+
const makeUploadDiff = ({ stats }) => {
|
|
34
|
+
if (!stats) {
|
|
35
|
+
return null;
|
|
36
|
+
}
|
|
37
|
+
if (stats.addedFiles === 0 && stats.removedFiles === 0) {
|
|
38
|
+
return cli_1.CliInternals.chalk.gray(`(Unchanged)`);
|
|
39
|
+
}
|
|
40
|
+
const total = stats.addedFiles + stats.removedFiles;
|
|
41
|
+
return cli_1.CliInternals.chalk.gray(`(${[
|
|
42
|
+
stats.addedFiles ? `+${stats.addedFiles}` : null,
|
|
43
|
+
stats.removedFiles ? `-${stats.removedFiles}` : null,
|
|
44
|
+
]
|
|
45
|
+
.filter(remotion_1.Internals.truthy)
|
|
46
|
+
.join(',')} ${total === 1 ? 'file' : 'files'})`);
|
|
47
|
+
};
|
|
48
|
+
const makeDeployProgressBar = ({ sizeUploaded, totalSize, doneIn, stats, }) => {
|
|
34
49
|
const progress = totalSize === null ? 0 : sizeUploaded / totalSize;
|
|
35
50
|
return [
|
|
36
51
|
'☁️ ',
|
|
@@ -42,6 +57,7 @@ const makeDeployProgressBar = ({ sizeUploaded, totalSize, doneIn, }) => {
|
|
|
42
57
|
? `${cli_1.CliInternals.formatBytes(sizeUploaded)}/${cli_1.CliInternals.formatBytes(totalSize)}`
|
|
43
58
|
: ''
|
|
44
59
|
: cli_1.CliInternals.chalk.gray(`${doneIn}ms`),
|
|
60
|
+
makeUploadDiff({ stats }),
|
|
45
61
|
]
|
|
46
62
|
.filter(remotion_1.Internals.truthy)
|
|
47
63
|
.join(' ');
|
package/dist/cli/index.js
CHANGED
|
@@ -97,7 +97,7 @@ const matchCommand = (args, remotionRoot) => {
|
|
|
97
97
|
(0, quit_1.quit)(1);
|
|
98
98
|
};
|
|
99
99
|
const executeCommand = async (args, remotionRoot) => {
|
|
100
|
-
var _a, _b;
|
|
100
|
+
var _a, _b, _c;
|
|
101
101
|
try {
|
|
102
102
|
(0, is_cli_1.setIsCli)(true);
|
|
103
103
|
await matchCommand(args, remotionRoot);
|
|
@@ -131,6 +131,11 @@ AWS returned an "AccessDenied" error message meaning a permission is missing. Re
|
|
|
131
131
|
log_1.Log.error(`
|
|
132
132
|
AWS returned an "TooManyRequestsException" error message which could mean you reached the concurrency limit of AWS Lambda. You can increase the limit - read this troubleshooting page: ${docs_url_1.DOCS_URL}/docs/lambda/troubleshooting/rate-limit. The original error message is:
|
|
133
133
|
`.trim());
|
|
134
|
+
}
|
|
135
|
+
if ((_c = error.stack) === null || _c === void 0 ? void 0 : _c.includes('The security token included in the request is invalid')) {
|
|
136
|
+
log_1.Log.error(`
|
|
137
|
+
AWS returned an error message "The security token included in the request is invalid". A possible reason for this is that you did not enable the region in your AWS account under "Account". The original message is:
|
|
138
|
+
`);
|
|
134
139
|
}
|
|
135
140
|
log_1.Log.error(error.stack);
|
|
136
141
|
(0, quit_1.quit)(1);
|
|
@@ -2,9 +2,10 @@
|
|
|
2
2
|
var _a;
|
|
3
3
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
4
4
|
exports.executablePath = void 0;
|
|
5
|
-
if (/^AWS_Lambda_nodejs(?:10|12|14)[.]x$/.test((_a = process.env.AWS_EXECUTION_ENV) !== null && _a !== void 0 ? _a : '') === true) {
|
|
5
|
+
if (/^AWS_Lambda_nodejs(?:10|12|14|16|18)[.]x$/.test((_a = process.env.AWS_EXECUTION_ENV) !== null && _a !== void 0 ? _a : '') === true) {
|
|
6
6
|
if (process.env.FONTCONFIG_PATH === undefined) {
|
|
7
7
|
process.env.FONTCONFIG_PATH = '/opt';
|
|
8
|
+
process.env.FONTCONFIG_FILE = '/opt/fonts.conf';
|
|
8
9
|
}
|
|
9
10
|
process.env.LD_LIBRARY_PATH = '/opt/lib:/opt/bin';
|
|
10
11
|
}
|
|
@@ -1 +1 @@
|
|
|
1
|
-
export declare const getCurrentRegionInFunction: () => "eu-central-1" | "eu-west-1" | "eu-west-2" | "us-east-1" | "us-east-2" | "us-west-2" | "ap-south-1" | "ap-southeast-1" | "ap-southeast-2" | "ap-northeast-1";
|
|
1
|
+
export declare const getCurrentRegionInFunction: () => "eu-central-1" | "eu-west-1" | "eu-west-2" | "eu-west-3" | "eu-south-1" | "eu-north-1" | "us-east-1" | "us-east-2" | "us-west-1" | "us-west-2" | "af-south-1" | "ap-south-1" | "ap-east-1" | "ap-southeast-1" | "ap-southeast-2" | "ap-northeast-1" | "ap-northeast-2" | "ap-northeast-3" | "ca-central-1" | "me-south-1" | "sa-east-1";
|
|
@@ -1,2 +1,2 @@
|
|
|
1
|
-
export declare const AWS_REGIONS: readonly ["eu-central-1", "eu-west-1", "eu-west-2", "us-east-1", "us-east-2", "us-west-2", "ap-south-1", "ap-southeast-1", "ap-southeast-2", "ap-northeast-1"];
|
|
1
|
+
export declare const AWS_REGIONS: readonly ["eu-central-1", "eu-west-1", "eu-west-2", "eu-west-3", "eu-south-1", "eu-north-1", "us-east-1", "us-east-2", "us-west-1", "us-west-2", "af-south-1", "ap-south-1", "ap-east-1", "ap-southeast-1", "ap-southeast-2", "ap-northeast-1", "ap-northeast-2", "ap-northeast-3", "ca-central-1", "me-south-1", "sa-east-1"];
|
|
2
2
|
export declare type AwsRegion = typeof AWS_REGIONS[number];
|
|
@@ -5,11 +5,22 @@ exports.AWS_REGIONS = [
|
|
|
5
5
|
'eu-central-1',
|
|
6
6
|
'eu-west-1',
|
|
7
7
|
'eu-west-2',
|
|
8
|
+
'eu-west-3',
|
|
9
|
+
'eu-south-1',
|
|
10
|
+
'eu-north-1',
|
|
8
11
|
'us-east-1',
|
|
9
12
|
'us-east-2',
|
|
13
|
+
'us-west-1',
|
|
10
14
|
'us-west-2',
|
|
15
|
+
'af-south-1',
|
|
11
16
|
'ap-south-1',
|
|
17
|
+
'ap-east-1',
|
|
12
18
|
'ap-southeast-1',
|
|
13
19
|
'ap-southeast-2',
|
|
14
20
|
'ap-northeast-1',
|
|
21
|
+
'ap-northeast-2',
|
|
22
|
+
'ap-northeast-3',
|
|
23
|
+
'ca-central-1',
|
|
24
|
+
'me-south-1',
|
|
25
|
+
'sa-east-1',
|
|
15
26
|
];
|