@aws-sdk/client-cleanrooms 3.1031.0 → 3.1033.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist-cjs/endpoint/bdd.js +49 -0
- package/dist-cjs/endpoint/endpointResolver.js +2 -2
- package/dist-cjs/schemas/schemas_0.js +2 -2
- package/dist-es/endpoint/bdd.js +46 -0
- package/dist-es/endpoint/endpointResolver.js +3 -3
- package/dist-es/schemas/schemas_0.js +2 -2
- package/dist-types/commands/GetProtectedJobCommand.d.ts +5 -0
- package/dist-types/commands/StartProtectedJobCommand.d.ts +10 -0
- package/dist-types/commands/UpdateProtectedJobCommand.d.ts +5 -0
- package/dist-types/endpoint/bdd.d.ts +2 -0
- package/dist-types/models/models_0.d.ts +39 -34
- package/dist-types/ts3.4/endpoint/bdd.d.ts +2 -0
- package/dist-types/ts3.4/models/models_0.d.ts +18 -17
- package/package.json +5 -5
- package/dist-cjs/endpoint/ruleset.js +0 -7
- package/dist-es/endpoint/ruleset.js +0 -4
- package/dist-types/endpoint/ruleset.d.ts +0 -2
- package/dist-types/ts3.4/endpoint/ruleset.d.ts +0 -2
|
@@ -0,0 +1,49 @@
|
|
|
1
|
+
"use strict";
|
|
2
|
+
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
+
exports.bdd = void 0;
|
|
4
|
+
const util_endpoints_1 = require("@smithy/util-endpoints");
|
|
5
|
+
const k = "ref";
|
|
6
|
+
const a = -1, b = true, c = "isSet", d = "PartitionResult", e = "booleanEquals", f = "getAttr", g = { [k]: "Endpoint" }, h = { [k]: d }, i = {}, j = [{ [k]: "Region" }];
|
|
7
|
+
const _data = {
|
|
8
|
+
conditions: [
|
|
9
|
+
[c, [g]],
|
|
10
|
+
[c, j],
|
|
11
|
+
["aws.partition", j, d],
|
|
12
|
+
[e, [{ [k]: "UseFIPS" }, b]],
|
|
13
|
+
[e, [{ [k]: "UseDualStack" }, b]],
|
|
14
|
+
[e, [{ fn: f, argv: [h, "supportsDualStack"] }, b]],
|
|
15
|
+
[e, [{ fn: f, argv: [h, "supportsFIPS"] }, b]]
|
|
16
|
+
],
|
|
17
|
+
results: [
|
|
18
|
+
[a],
|
|
19
|
+
[a, "Invalid Configuration: FIPS and custom endpoint are not supported"],
|
|
20
|
+
[a, "Invalid Configuration: Dualstack and custom endpoint are not supported"],
|
|
21
|
+
[g, i],
|
|
22
|
+
["https://cleanrooms-fips.{Region}.{PartitionResult#dualStackDnsSuffix}", i],
|
|
23
|
+
[a, "FIPS and DualStack are enabled, but this partition does not support one or both"],
|
|
24
|
+
["https://cleanrooms-fips.{Region}.{PartitionResult#dnsSuffix}", i],
|
|
25
|
+
[a, "FIPS is enabled but this partition does not support FIPS"],
|
|
26
|
+
["https://cleanrooms.{Region}.{PartitionResult#dualStackDnsSuffix}", i],
|
|
27
|
+
[a, "DualStack is enabled but this partition does not support DualStack"],
|
|
28
|
+
["https://cleanrooms.{Region}.{PartitionResult#dnsSuffix}", i],
|
|
29
|
+
[a, "Invalid Configuration: Missing Region"]
|
|
30
|
+
]
|
|
31
|
+
};
|
|
32
|
+
const root = 2;
|
|
33
|
+
const r = 100_000_000;
|
|
34
|
+
const nodes = new Int32Array([
|
|
35
|
+
-1, 1, -1,
|
|
36
|
+
0, 12, 3,
|
|
37
|
+
1, 4, r + 11,
|
|
38
|
+
2, 5, r + 11,
|
|
39
|
+
3, 8, 6,
|
|
40
|
+
4, 7, r + 10,
|
|
41
|
+
5, r + 8, r + 9,
|
|
42
|
+
4, 10, 9,
|
|
43
|
+
6, r + 6, r + 7,
|
|
44
|
+
5, 11, r + 5,
|
|
45
|
+
6, r + 4, r + 5,
|
|
46
|
+
3, r + 1, 13,
|
|
47
|
+
4, r + 2, r + 3,
|
|
48
|
+
]);
|
|
49
|
+
exports.bdd = util_endpoints_1.BinaryDecisionDiagram.from(nodes, root, _data.conditions, _data.results);
|
|
@@ -3,13 +3,13 @@ Object.defineProperty(exports, "__esModule", { value: true });
|
|
|
3
3
|
exports.defaultEndpointResolver = void 0;
|
|
4
4
|
const util_endpoints_1 = require("@aws-sdk/util-endpoints");
|
|
5
5
|
const util_endpoints_2 = require("@smithy/util-endpoints");
|
|
6
|
-
const
|
|
6
|
+
const bdd_1 = require("./bdd");
|
|
7
7
|
const cache = new util_endpoints_2.EndpointCache({
|
|
8
8
|
size: 50,
|
|
9
9
|
params: ["Endpoint", "Region", "UseDualStack", "UseFIPS"],
|
|
10
10
|
});
|
|
11
11
|
const defaultEndpointResolver = (endpointParams, context = {}) => {
|
|
12
|
-
return cache.get(endpointParams, () => (0, util_endpoints_2.
|
|
12
|
+
return cache.get(endpointParams, () => (0, util_endpoints_2.decideEndpoint)(bdd_1.bdd, {
|
|
13
13
|
endpointParams: endpointParams,
|
|
14
14
|
logger: context.logger,
|
|
15
15
|
}));
|
|
@@ -2235,8 +2235,8 @@ exports.ProtectedJobSummary$ = [3, n0, _PJSr,
|
|
|
2235
2235
|
];
|
|
2236
2236
|
exports.ProtectedJobWorkerComputeConfiguration$ = [3, n0, _PJWCC,
|
|
2237
2237
|
0,
|
|
2238
|
-
[_t, _nu],
|
|
2239
|
-
[0, 1], 2
|
|
2238
|
+
[_t, _nu, _pr],
|
|
2239
|
+
[0, 1, () => exports.WorkerComputeConfigurationProperties$], 2
|
|
2240
2240
|
];
|
|
2241
2241
|
exports.ProtectedQuery$ = [3, n0, _PQ,
|
|
2242
2242
|
0,
|
|
@@ -0,0 +1,46 @@
|
|
|
1
|
+
import { BinaryDecisionDiagram } from "@smithy/util-endpoints";
|
|
2
|
+
const k = "ref";
|
|
3
|
+
const a = -1, b = true, c = "isSet", d = "PartitionResult", e = "booleanEquals", f = "getAttr", g = { [k]: "Endpoint" }, h = { [k]: d }, i = {}, j = [{ [k]: "Region" }];
|
|
4
|
+
const _data = {
|
|
5
|
+
conditions: [
|
|
6
|
+
[c, [g]],
|
|
7
|
+
[c, j],
|
|
8
|
+
["aws.partition", j, d],
|
|
9
|
+
[e, [{ [k]: "UseFIPS" }, b]],
|
|
10
|
+
[e, [{ [k]: "UseDualStack" }, b]],
|
|
11
|
+
[e, [{ fn: f, argv: [h, "supportsDualStack"] }, b]],
|
|
12
|
+
[e, [{ fn: f, argv: [h, "supportsFIPS"] }, b]]
|
|
13
|
+
],
|
|
14
|
+
results: [
|
|
15
|
+
[a],
|
|
16
|
+
[a, "Invalid Configuration: FIPS and custom endpoint are not supported"],
|
|
17
|
+
[a, "Invalid Configuration: Dualstack and custom endpoint are not supported"],
|
|
18
|
+
[g, i],
|
|
19
|
+
["https://cleanrooms-fips.{Region}.{PartitionResult#dualStackDnsSuffix}", i],
|
|
20
|
+
[a, "FIPS and DualStack are enabled, but this partition does not support one or both"],
|
|
21
|
+
["https://cleanrooms-fips.{Region}.{PartitionResult#dnsSuffix}", i],
|
|
22
|
+
[a, "FIPS is enabled but this partition does not support FIPS"],
|
|
23
|
+
["https://cleanrooms.{Region}.{PartitionResult#dualStackDnsSuffix}", i],
|
|
24
|
+
[a, "DualStack is enabled but this partition does not support DualStack"],
|
|
25
|
+
["https://cleanrooms.{Region}.{PartitionResult#dnsSuffix}", i],
|
|
26
|
+
[a, "Invalid Configuration: Missing Region"]
|
|
27
|
+
]
|
|
28
|
+
};
|
|
29
|
+
const root = 2;
|
|
30
|
+
const r = 100_000_000;
|
|
31
|
+
const nodes = new Int32Array([
|
|
32
|
+
-1, 1, -1,
|
|
33
|
+
0, 12, 3,
|
|
34
|
+
1, 4, r + 11,
|
|
35
|
+
2, 5, r + 11,
|
|
36
|
+
3, 8, 6,
|
|
37
|
+
4, 7, r + 10,
|
|
38
|
+
5, r + 8, r + 9,
|
|
39
|
+
4, 10, 9,
|
|
40
|
+
6, r + 6, r + 7,
|
|
41
|
+
5, 11, r + 5,
|
|
42
|
+
6, r + 4, r + 5,
|
|
43
|
+
3, r + 1, 13,
|
|
44
|
+
4, r + 2, r + 3,
|
|
45
|
+
]);
|
|
46
|
+
export const bdd = BinaryDecisionDiagram.from(nodes, root, _data.conditions, _data.results);
|
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
import { awsEndpointFunctions } from "@aws-sdk/util-endpoints";
|
|
2
|
-
import { customEndpointFunctions,
|
|
3
|
-
import {
|
|
2
|
+
import { customEndpointFunctions, decideEndpoint, EndpointCache } from "@smithy/util-endpoints";
|
|
3
|
+
import { bdd } from "./bdd";
|
|
4
4
|
const cache = new EndpointCache({
|
|
5
5
|
size: 50,
|
|
6
6
|
params: ["Endpoint", "Region", "UseDualStack", "UseFIPS"],
|
|
7
7
|
});
|
|
8
8
|
export const defaultEndpointResolver = (endpointParams, context = {}) => {
|
|
9
|
-
return cache.get(endpointParams, () =>
|
|
9
|
+
return cache.get(endpointParams, () => decideEndpoint(bdd, {
|
|
10
10
|
endpointParams: endpointParams,
|
|
11
11
|
logger: context.logger,
|
|
12
12
|
}));
|
|
@@ -2223,8 +2223,8 @@ export var ProtectedJobSummary$ = [3, n0, _PJSr,
|
|
|
2223
2223
|
];
|
|
2224
2224
|
export var ProtectedJobWorkerComputeConfiguration$ = [3, n0, _PJWCC,
|
|
2225
2225
|
0,
|
|
2226
|
-
[_t, _nu],
|
|
2227
|
-
[0, 1], 2
|
|
2226
|
+
[_t, _nu, _pr],
|
|
2227
|
+
[0, 1, () => WorkerComputeConfigurationProperties$], 2
|
|
2228
2228
|
];
|
|
2229
2229
|
export var ProtectedQuery$ = [3, n0, _PQ,
|
|
2230
2230
|
0,
|
|
@@ -92,6 +92,11 @@ declare const GetProtectedJobCommand_base: {
|
|
|
92
92
|
* // worker: { // ProtectedJobWorkerComputeConfiguration
|
|
93
93
|
* // type: "CR.1X" || "CR.4X", // required
|
|
94
94
|
* // number: Number("int"), // required
|
|
95
|
+
* // properties: { // WorkerComputeConfigurationProperties Union: only one key present
|
|
96
|
+
* // spark: { // SparkProperties
|
|
97
|
+
* // "<keys>": "STRING_VALUE",
|
|
98
|
+
* // },
|
|
99
|
+
* // },
|
|
95
100
|
* // },
|
|
96
101
|
* // },
|
|
97
102
|
* // },
|
|
@@ -56,6 +56,11 @@ declare const StartProtectedJobCommand_base: {
|
|
|
56
56
|
* worker: { // ProtectedJobWorkerComputeConfiguration
|
|
57
57
|
* type: "CR.1X" || "CR.4X", // required
|
|
58
58
|
* number: Number("int"), // required
|
|
59
|
+
* properties: { // WorkerComputeConfigurationProperties Union: only one key present
|
|
60
|
+
* spark: { // SparkProperties
|
|
61
|
+
* "<keys>": "STRING_VALUE",
|
|
62
|
+
* },
|
|
63
|
+
* },
|
|
59
64
|
* },
|
|
60
65
|
* },
|
|
61
66
|
* };
|
|
@@ -111,6 +116,11 @@ declare const StartProtectedJobCommand_base: {
|
|
|
111
116
|
* // worker: { // ProtectedJobWorkerComputeConfiguration
|
|
112
117
|
* // type: "CR.1X" || "CR.4X", // required
|
|
113
118
|
* // number: Number("int"), // required
|
|
119
|
+
* // properties: { // WorkerComputeConfigurationProperties Union: only one key present
|
|
120
|
+
* // spark: { // SparkProperties
|
|
121
|
+
* // "<keys>": "STRING_VALUE",
|
|
122
|
+
* // },
|
|
123
|
+
* // },
|
|
114
124
|
* // },
|
|
115
125
|
* // },
|
|
116
126
|
* // },
|
|
@@ -93,6 +93,11 @@ declare const UpdateProtectedJobCommand_base: {
|
|
|
93
93
|
* // worker: { // ProtectedJobWorkerComputeConfiguration
|
|
94
94
|
* // type: "CR.1X" || "CR.4X", // required
|
|
95
95
|
* // number: Number("int"), // required
|
|
96
|
+
* // properties: { // WorkerComputeConfigurationProperties Union: only one key present
|
|
97
|
+
* // spark: { // SparkProperties
|
|
98
|
+
* // "<keys>": "STRING_VALUE",
|
|
99
|
+
* // },
|
|
100
|
+
* // },
|
|
96
101
|
* // },
|
|
97
102
|
* // },
|
|
98
103
|
* // },
|
|
@@ -2423,7 +2423,7 @@ export interface MemberChangeSpecification {
|
|
|
2423
2423
|
*/
|
|
2424
2424
|
accountId: string | undefined;
|
|
2425
2425
|
/**
|
|
2426
|
-
* <p>The abilities granted to the collaboration member. These determine what actions the member can perform within the collaboration.</p> <note> <p>The following values are currently not supported: <code>CAN_QUERY</code
|
|
2426
|
+
* <p>The abilities granted to the collaboration member. These determine what actions the member can perform within the collaboration.</p> <note> <p>The following values are currently not supported: <code>CAN_QUERY</code> and <code>CAN_RUN_JOB</code>. </p> <p>Set the value of <code>memberAbilities</code> to <code>[]</code> to allow a member to contribute data.</p> <p>Set the value of <code>memberAbilities</code> to <code>[CAN_RECEIVE_RESULTS]</code> to allow a member to contribute data and receive results.</p> </note>
|
|
2427
2427
|
* @public
|
|
2428
2428
|
*/
|
|
2429
2429
|
memberAbilities: MemberAbility[] | undefined;
|
|
@@ -6601,6 +6601,39 @@ export interface GetProtectedJobInput {
|
|
|
6601
6601
|
*/
|
|
6602
6602
|
protectedJobIdentifier: string | undefined;
|
|
6603
6603
|
}
|
|
6604
|
+
/**
|
|
6605
|
+
* <p>The configuration properties that define the compute environment settings for workers in Clean Rooms. These properties enable customization of the underlying compute environment to optimize performance for your specific workloads.</p>
|
|
6606
|
+
* @public
|
|
6607
|
+
*/
|
|
6608
|
+
export type WorkerComputeConfigurationProperties = WorkerComputeConfigurationProperties.SparkMember | WorkerComputeConfigurationProperties.$UnknownMember;
|
|
6609
|
+
/**
|
|
6610
|
+
* @public
|
|
6611
|
+
*/
|
|
6612
|
+
export declare namespace WorkerComputeConfigurationProperties {
|
|
6613
|
+
/**
|
|
6614
|
+
* <p>The Spark configuration properties for SQL and PySpark workloads. This map contains key-value pairs that configure Apache Spark settings to optimize performance for your data processing jobs. You can specify up to 50 Spark properties, with each key being 1-200 characters and each value being 0-500 characters. These properties allow you to adjust compute capacity for large datasets and complex workloads.</p>
|
|
6615
|
+
* @public
|
|
6616
|
+
*/
|
|
6617
|
+
interface SparkMember {
|
|
6618
|
+
spark: Record<string, string>;
|
|
6619
|
+
$unknown?: never;
|
|
6620
|
+
}
|
|
6621
|
+
/**
|
|
6622
|
+
* @public
|
|
6623
|
+
*/
|
|
6624
|
+
interface $UnknownMember {
|
|
6625
|
+
spark?: never;
|
|
6626
|
+
$unknown: [string, any];
|
|
6627
|
+
}
|
|
6628
|
+
/**
|
|
6629
|
+
* @deprecated unused in schema-serde mode.
|
|
6630
|
+
*
|
|
6631
|
+
*/
|
|
6632
|
+
interface Visitor<T> {
|
|
6633
|
+
spark: (value: Record<string, string>) => T;
|
|
6634
|
+
_: (name: string, value: any) => T;
|
|
6635
|
+
}
|
|
6636
|
+
}
|
|
6604
6637
|
/**
|
|
6605
6638
|
* <p>The configuration of the compute resources for a PySpark job.</p>
|
|
6606
6639
|
* @public
|
|
@@ -6616,6 +6649,11 @@ export interface ProtectedJobWorkerComputeConfiguration {
|
|
|
6616
6649
|
* @public
|
|
6617
6650
|
*/
|
|
6618
6651
|
number: number | undefined;
|
|
6652
|
+
/**
|
|
6653
|
+
* <p>The configuration properties for the worker compute environment. These properties allow you to customize the compute settings for your Clean Rooms workloads.</p>
|
|
6654
|
+
* @public
|
|
6655
|
+
*/
|
|
6656
|
+
properties?: WorkerComputeConfigurationProperties | undefined;
|
|
6619
6657
|
}
|
|
6620
6658
|
/**
|
|
6621
6659
|
* <p>The configuration of the compute resources for a PySpark job.</p>
|
|
@@ -6956,39 +6994,6 @@ export interface GetProtectedQueryInput {
|
|
|
6956
6994
|
*/
|
|
6957
6995
|
protectedQueryIdentifier: string | undefined;
|
|
6958
6996
|
}
|
|
6959
|
-
/**
|
|
6960
|
-
* <p>The configuration properties that define the compute environment settings for workers in Clean Rooms. These properties enable customization of the underlying compute environment to optimize performance for your specific workloads.</p>
|
|
6961
|
-
* @public
|
|
6962
|
-
*/
|
|
6963
|
-
export type WorkerComputeConfigurationProperties = WorkerComputeConfigurationProperties.SparkMember | WorkerComputeConfigurationProperties.$UnknownMember;
|
|
6964
|
-
/**
|
|
6965
|
-
* @public
|
|
6966
|
-
*/
|
|
6967
|
-
export declare namespace WorkerComputeConfigurationProperties {
|
|
6968
|
-
/**
|
|
6969
|
-
* <p>The Spark configuration properties for SQL workloads. This map contains key-value pairs that configure Apache Spark settings to optimize performance for your data processing jobs. You can specify up to 50 Spark properties, with each key being 1-200 characters and each value being 0-500 characters. These properties allow you to adjust compute capacity for large datasets and complex workloads.</p>
|
|
6970
|
-
* @public
|
|
6971
|
-
*/
|
|
6972
|
-
interface SparkMember {
|
|
6973
|
-
spark: Record<string, string>;
|
|
6974
|
-
$unknown?: never;
|
|
6975
|
-
}
|
|
6976
|
-
/**
|
|
6977
|
-
* @public
|
|
6978
|
-
*/
|
|
6979
|
-
interface $UnknownMember {
|
|
6980
|
-
spark?: never;
|
|
6981
|
-
$unknown: [string, any];
|
|
6982
|
-
}
|
|
6983
|
-
/**
|
|
6984
|
-
* @deprecated unused in schema-serde mode.
|
|
6985
|
-
*
|
|
6986
|
-
*/
|
|
6987
|
-
interface Visitor<T> {
|
|
6988
|
-
spark: (value: Record<string, string>) => T;
|
|
6989
|
-
_: (name: string, value: any) => T;
|
|
6990
|
-
}
|
|
6991
|
-
}
|
|
6992
6997
|
/**
|
|
6993
6998
|
* <p> The configuration of the compute resources for workers running an analysis with the Clean Rooms SQL analytics engine.</p>
|
|
6994
6999
|
* @public
|
|
@@ -1962,9 +1962,27 @@ export interface GetProtectedJobInput {
|
|
|
1962
1962
|
membershipIdentifier: string | undefined;
|
|
1963
1963
|
protectedJobIdentifier: string | undefined;
|
|
1964
1964
|
}
|
|
1965
|
+
export type WorkerComputeConfigurationProperties =
|
|
1966
|
+
| WorkerComputeConfigurationProperties.SparkMember
|
|
1967
|
+
| WorkerComputeConfigurationProperties.$UnknownMember;
|
|
1968
|
+
export declare namespace WorkerComputeConfigurationProperties {
|
|
1969
|
+
interface SparkMember {
|
|
1970
|
+
spark: Record<string, string>;
|
|
1971
|
+
$unknown?: never;
|
|
1972
|
+
}
|
|
1973
|
+
interface $UnknownMember {
|
|
1974
|
+
spark?: never;
|
|
1975
|
+
$unknown: [string, any];
|
|
1976
|
+
}
|
|
1977
|
+
interface Visitor<T> {
|
|
1978
|
+
spark: (value: Record<string, string>) => T;
|
|
1979
|
+
_: (name: string, value: any) => T;
|
|
1980
|
+
}
|
|
1981
|
+
}
|
|
1965
1982
|
export interface ProtectedJobWorkerComputeConfiguration {
|
|
1966
1983
|
type: ProtectedJobWorkerComputeType | undefined;
|
|
1967
1984
|
number: number | undefined;
|
|
1985
|
+
properties?: WorkerComputeConfigurationProperties | undefined;
|
|
1968
1986
|
}
|
|
1969
1987
|
export type ProtectedJobComputeConfiguration =
|
|
1970
1988
|
| ProtectedJobComputeConfiguration.WorkerMember
|
|
@@ -2089,23 +2107,6 @@ export interface GetProtectedQueryInput {
|
|
|
2089
2107
|
membershipIdentifier: string | undefined;
|
|
2090
2108
|
protectedQueryIdentifier: string | undefined;
|
|
2091
2109
|
}
|
|
2092
|
-
export type WorkerComputeConfigurationProperties =
|
|
2093
|
-
| WorkerComputeConfigurationProperties.SparkMember
|
|
2094
|
-
| WorkerComputeConfigurationProperties.$UnknownMember;
|
|
2095
|
-
export declare namespace WorkerComputeConfigurationProperties {
|
|
2096
|
-
interface SparkMember {
|
|
2097
|
-
spark: Record<string, string>;
|
|
2098
|
-
$unknown?: never;
|
|
2099
|
-
}
|
|
2100
|
-
interface $UnknownMember {
|
|
2101
|
-
spark?: never;
|
|
2102
|
-
$unknown: [string, any];
|
|
2103
|
-
}
|
|
2104
|
-
interface Visitor<T> {
|
|
2105
|
-
spark: (value: Record<string, string>) => T;
|
|
2106
|
-
_: (name: string, value: any) => T;
|
|
2107
|
-
}
|
|
2108
|
-
}
|
|
2109
2110
|
export interface WorkerComputeConfiguration {
|
|
2110
2111
|
type?: WorkerComputeType | undefined;
|
|
2111
2112
|
number?: number | undefined;
|
package/package.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@aws-sdk/client-cleanrooms",
|
|
3
3
|
"description": "AWS SDK for JavaScript Cleanrooms Client for Node.js, Browser and React Native",
|
|
4
|
-
"version": "3.
|
|
4
|
+
"version": "3.1033.0",
|
|
5
5
|
"scripts": {
|
|
6
6
|
"build": "concurrently 'yarn:build:types' 'yarn:build:es' && yarn build:cjs",
|
|
7
7
|
"build:cjs": "node ../../scripts/compilation/inline client-cleanrooms",
|
|
@@ -21,17 +21,17 @@
|
|
|
21
21
|
"dependencies": {
|
|
22
22
|
"@aws-crypto/sha256-browser": "5.2.0",
|
|
23
23
|
"@aws-crypto/sha256-js": "5.2.0",
|
|
24
|
-
"@aws-sdk/core": "^3.974.
|
|
25
|
-
"@aws-sdk/credential-provider-node": "^3.972.
|
|
24
|
+
"@aws-sdk/core": "^3.974.2",
|
|
25
|
+
"@aws-sdk/credential-provider-node": "^3.972.33",
|
|
26
26
|
"@aws-sdk/middleware-host-header": "^3.972.10",
|
|
27
27
|
"@aws-sdk/middleware-logger": "^3.972.10",
|
|
28
28
|
"@aws-sdk/middleware-recursion-detection": "^3.972.11",
|
|
29
|
-
"@aws-sdk/middleware-user-agent": "^3.972.
|
|
29
|
+
"@aws-sdk/middleware-user-agent": "^3.972.32",
|
|
30
30
|
"@aws-sdk/region-config-resolver": "^3.972.12",
|
|
31
31
|
"@aws-sdk/types": "^3.973.8",
|
|
32
32
|
"@aws-sdk/util-endpoints": "^3.996.7",
|
|
33
33
|
"@aws-sdk/util-user-agent-browser": "^3.972.10",
|
|
34
|
-
"@aws-sdk/util-user-agent-node": "^3.973.
|
|
34
|
+
"@aws-sdk/util-user-agent-node": "^3.973.18",
|
|
35
35
|
"@smithy/config-resolver": "^4.4.16",
|
|
36
36
|
"@smithy/core": "^3.23.15",
|
|
37
37
|
"@smithy/fetch-http-handler": "^5.3.17",
|
|
@@ -1,7 +0,0 @@
|
|
|
1
|
-
"use strict";
|
|
2
|
-
Object.defineProperty(exports, "__esModule", { value: true });
|
|
3
|
-
exports.ruleSet = void 0;
|
|
4
|
-
const s = "required", t = "fn", u = "argv", v = "ref";
|
|
5
|
-
const a = true, b = "isSet", c = "booleanEquals", d = "error", e = "endpoint", f = "tree", g = "PartitionResult", h = { [s]: false, "type": "string" }, i = { [s]: true, "default": false, "type": "boolean" }, j = { [v]: "Endpoint" }, k = { [t]: c, [u]: [{ [v]: "UseFIPS" }, true] }, l = { [t]: c, [u]: [{ [v]: "UseDualStack" }, true] }, m = {}, n = { [t]: "getAttr", [u]: [{ [v]: g }, "supportsFIPS"] }, o = { [t]: c, [u]: [true, { [t]: "getAttr", [u]: [{ [v]: g }, "supportsDualStack"] }] }, p = [k], q = [l], r = [{ [v]: "Region" }];
|
|
6
|
-
const _data = { version: "1.0", parameters: { Region: h, UseDualStack: i, UseFIPS: i, Endpoint: h }, rules: [{ conditions: [{ [t]: b, [u]: [j] }], rules: [{ conditions: p, error: "Invalid Configuration: FIPS and custom endpoint are not supported", type: d }, { rules: [{ conditions: q, error: "Invalid Configuration: Dualstack and custom endpoint are not supported", type: d }, { endpoint: { url: j, properties: m, headers: m }, type: e }], type: f }], type: f }, { rules: [{ conditions: [{ [t]: b, [u]: r }], rules: [{ conditions: [{ [t]: "aws.partition", [u]: r, assign: g }], rules: [{ conditions: [k, l], rules: [{ conditions: [{ [t]: c, [u]: [a, n] }, o], rules: [{ rules: [{ endpoint: { url: "https://cleanrooms-fips.{Region}.{PartitionResult#dualStackDnsSuffix}", properties: m, headers: m }, type: e }], type: f }], type: f }, { error: "FIPS and DualStack are enabled, but this partition does not support one or both", type: d }], type: f }, { conditions: p, rules: [{ conditions: [{ [t]: c, [u]: [n, a] }], rules: [{ rules: [{ endpoint: { url: "https://cleanrooms-fips.{Region}.{PartitionResult#dnsSuffix}", properties: m, headers: m }, type: e }], type: f }], type: f }, { error: "FIPS is enabled but this partition does not support FIPS", type: d }], type: f }, { conditions: q, rules: [{ conditions: [o], rules: [{ rules: [{ endpoint: { url: "https://cleanrooms.{Region}.{PartitionResult#dualStackDnsSuffix}", properties: m, headers: m }, type: e }], type: f }], type: f }, { error: "DualStack is enabled but this partition does not support DualStack", type: d }], type: f }, { rules: [{ endpoint: { url: "https://cleanrooms.{Region}.{PartitionResult#dnsSuffix}", properties: m, headers: m }, type: e }], type: f }], type: f }], type: f }, { error: "Invalid Configuration: Missing Region", type: d }], type: f }] };
|
|
7
|
-
exports.ruleSet = _data;
|
|
@@ -1,4 +0,0 @@
|
|
|
1
|
-
const s = "required", t = "fn", u = "argv", v = "ref";
|
|
2
|
-
const a = true, b = "isSet", c = "booleanEquals", d = "error", e = "endpoint", f = "tree", g = "PartitionResult", h = { [s]: false, "type": "string" }, i = { [s]: true, "default": false, "type": "boolean" }, j = { [v]: "Endpoint" }, k = { [t]: c, [u]: [{ [v]: "UseFIPS" }, true] }, l = { [t]: c, [u]: [{ [v]: "UseDualStack" }, true] }, m = {}, n = { [t]: "getAttr", [u]: [{ [v]: g }, "supportsFIPS"] }, o = { [t]: c, [u]: [true, { [t]: "getAttr", [u]: [{ [v]: g }, "supportsDualStack"] }] }, p = [k], q = [l], r = [{ [v]: "Region" }];
|
|
3
|
-
const _data = { version: "1.0", parameters: { Region: h, UseDualStack: i, UseFIPS: i, Endpoint: h }, rules: [{ conditions: [{ [t]: b, [u]: [j] }], rules: [{ conditions: p, error: "Invalid Configuration: FIPS and custom endpoint are not supported", type: d }, { rules: [{ conditions: q, error: "Invalid Configuration: Dualstack and custom endpoint are not supported", type: d }, { endpoint: { url: j, properties: m, headers: m }, type: e }], type: f }], type: f }, { rules: [{ conditions: [{ [t]: b, [u]: r }], rules: [{ conditions: [{ [t]: "aws.partition", [u]: r, assign: g }], rules: [{ conditions: [k, l], rules: [{ conditions: [{ [t]: c, [u]: [a, n] }, o], rules: [{ rules: [{ endpoint: { url: "https://cleanrooms-fips.{Region}.{PartitionResult#dualStackDnsSuffix}", properties: m, headers: m }, type: e }], type: f }], type: f }, { error: "FIPS and DualStack are enabled, but this partition does not support one or both", type: d }], type: f }, { conditions: p, rules: [{ conditions: [{ [t]: c, [u]: [n, a] }], rules: [{ rules: [{ endpoint: { url: "https://cleanrooms-fips.{Region}.{PartitionResult#dnsSuffix}", properties: m, headers: m }, type: e }], type: f }], type: f }, { error: "FIPS is enabled but this partition does not support FIPS", type: d }], type: f }, { conditions: q, rules: [{ conditions: [o], rules: [{ rules: [{ endpoint: { url: "https://cleanrooms.{Region}.{PartitionResult#dualStackDnsSuffix}", properties: m, headers: m }, type: e }], type: f }], type: f }, { error: "DualStack is enabled but this partition does not support DualStack", type: d }], type: f }, { rules: [{ endpoint: { url: "https://cleanrooms.{Region}.{PartitionResult#dnsSuffix}", properties: m, headers: m }, type: e }], type: f }], type: f }], type: f }, { error: "Invalid Configuration: Missing Region", type: d }], type: f }] };
|
|
4
|
-
export const ruleSet = _data;
|