@aws-sdk/client-dataexchange 3.301.0 → 3.303.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -16,45 +16,38 @@ class AccessDeniedException extends DataExchangeServiceException_1.DataExchangeS
16
16
  }
17
17
  }
18
18
  exports.AccessDeniedException = AccessDeniedException;
19
- var ServerSideEncryptionTypes;
20
- (function (ServerSideEncryptionTypes) {
21
- ServerSideEncryptionTypes["AES256"] = "AES256";
22
- ServerSideEncryptionTypes["aws_kms"] = "aws:kms";
23
- })(ServerSideEncryptionTypes = exports.ServerSideEncryptionTypes || (exports.ServerSideEncryptionTypes = {}));
24
- var ProtocolType;
25
- (function (ProtocolType) {
26
- ProtocolType["REST"] = "REST";
27
- })(ProtocolType = exports.ProtocolType || (exports.ProtocolType = {}));
28
- var LFResourceType;
29
- (function (LFResourceType) {
30
- LFResourceType["DATABASE"] = "DATABASE";
31
- LFResourceType["TABLE"] = "TABLE";
32
- })(LFResourceType = exports.LFResourceType || (exports.LFResourceType = {}));
33
- var LakeFormationDataPermissionType;
34
- (function (LakeFormationDataPermissionType) {
35
- LakeFormationDataPermissionType["LFTagPolicy"] = "LFTagPolicy";
36
- })(LakeFormationDataPermissionType = exports.LakeFormationDataPermissionType || (exports.LakeFormationDataPermissionType = {}));
37
- var LFPermission;
38
- (function (LFPermission) {
39
- LFPermission["DESCRIBE"] = "DESCRIBE";
40
- LFPermission["SELECT"] = "SELECT";
41
- })(LFPermission = exports.LFPermission || (exports.LFPermission = {}));
42
- var AssetType;
43
- (function (AssetType) {
44
- AssetType["API_GATEWAY_API"] = "API_GATEWAY_API";
45
- AssetType["LAKE_FORMATION_DATA_PERMISSION"] = "LAKE_FORMATION_DATA_PERMISSION";
46
- AssetType["REDSHIFT_DATA_SHARE"] = "REDSHIFT_DATA_SHARE";
47
- AssetType["S3_DATA_ACCESS"] = "S3_DATA_ACCESS";
48
- AssetType["S3_SNAPSHOT"] = "S3_SNAPSHOT";
49
- })(AssetType = exports.AssetType || (exports.AssetType = {}));
50
- var ResourceType;
51
- (function (ResourceType) {
52
- ResourceType["ASSET"] = "ASSET";
53
- ResourceType["DATA_SET"] = "DATA_SET";
54
- ResourceType["EVENT_ACTION"] = "EVENT_ACTION";
55
- ResourceType["JOB"] = "JOB";
56
- ResourceType["REVISION"] = "REVISION";
57
- })(ResourceType = exports.ResourceType || (exports.ResourceType = {}));
19
+ exports.ServerSideEncryptionTypes = {
20
+ AES256: "AES256",
21
+ aws_kms: "aws:kms",
22
+ };
23
+ exports.ProtocolType = {
24
+ REST: "REST",
25
+ };
26
+ exports.LFResourceType = {
27
+ DATABASE: "DATABASE",
28
+ TABLE: "TABLE",
29
+ };
30
+ exports.LakeFormationDataPermissionType = {
31
+ LFTagPolicy: "LFTagPolicy",
32
+ };
33
+ exports.LFPermission = {
34
+ DESCRIBE: "DESCRIBE",
35
+ SELECT: "SELECT",
36
+ };
37
+ exports.AssetType = {
38
+ API_GATEWAY_API: "API_GATEWAY_API",
39
+ LAKE_FORMATION_DATA_PERMISSION: "LAKE_FORMATION_DATA_PERMISSION",
40
+ REDSHIFT_DATA_SHARE: "REDSHIFT_DATA_SHARE",
41
+ S3_DATA_ACCESS: "S3_DATA_ACCESS",
42
+ S3_SNAPSHOT: "S3_SNAPSHOT",
43
+ };
44
+ exports.ResourceType = {
45
+ ASSET: "ASSET",
46
+ DATA_SET: "DATA_SET",
47
+ EVENT_ACTION: "EVENT_ACTION",
48
+ JOB: "JOB",
49
+ REVISION: "REVISION",
50
+ };
58
51
  class ConflictException extends DataExchangeServiceException_1.DataExchangeServiceException {
59
52
  constructor(opts) {
60
53
  super({
@@ -115,11 +108,10 @@ class ThrottlingException extends DataExchangeServiceException_1.DataExchangeSer
115
108
  }
116
109
  }
117
110
  exports.ThrottlingException = ThrottlingException;
118
- var ExceptionCause;
119
- (function (ExceptionCause) {
120
- ExceptionCause["InsufficientS3BucketPolicy"] = "InsufficientS3BucketPolicy";
121
- ExceptionCause["S3AccessDenied"] = "S3AccessDenied";
122
- })(ExceptionCause = exports.ExceptionCause || (exports.ExceptionCause = {}));
111
+ exports.ExceptionCause = {
112
+ InsufficientS3BucketPolicy: "InsufficientS3BucketPolicy",
113
+ S3AccessDenied: "S3AccessDenied",
114
+ };
123
115
  class ValidationException extends DataExchangeServiceException_1.DataExchangeServiceException {
124
116
  constructor(opts) {
125
117
  super({
@@ -135,52 +127,49 @@ class ValidationException extends DataExchangeServiceException_1.DataExchangeSer
135
127
  }
136
128
  }
137
129
  exports.ValidationException = ValidationException;
138
- var Code;
139
- (function (Code) {
140
- Code["ACCESS_DENIED_EXCEPTION"] = "ACCESS_DENIED_EXCEPTION";
141
- Code["INTERNAL_SERVER_EXCEPTION"] = "INTERNAL_SERVER_EXCEPTION";
142
- Code["MALWARE_DETECTED"] = "MALWARE_DETECTED";
143
- Code["MALWARE_SCAN_ENCRYPTED_FILE"] = "MALWARE_SCAN_ENCRYPTED_FILE";
144
- Code["RESOURCE_NOT_FOUND_EXCEPTION"] = "RESOURCE_NOT_FOUND_EXCEPTION";
145
- Code["SERVICE_QUOTA_EXCEEDED_EXCEPTION"] = "SERVICE_QUOTA_EXCEEDED_EXCEPTION";
146
- Code["VALIDATION_EXCEPTION"] = "VALIDATION_EXCEPTION";
147
- })(Code = exports.Code || (exports.Code = {}));
148
- var Origin;
149
- (function (Origin) {
150
- Origin["ENTITLED"] = "ENTITLED";
151
- Origin["OWNED"] = "OWNED";
152
- })(Origin = exports.Origin || (exports.Origin = {}));
153
- var LimitName;
154
- (function (LimitName) {
155
- LimitName["AWS_Lake_Formation_data_permission_assets_per_revision"] = "AWS Lake Formation data permission assets per revision";
156
- LimitName["Amazon_API_Gateway_API_assets_per_revision"] = "Amazon API Gateway API assets per revision";
157
- LimitName["Amazon_Redshift_datashare_assets_per_import_job_from_Redshift"] = "Amazon Redshift datashare assets per import job from Redshift";
158
- LimitName["Amazon_Redshift_datashare_assets_per_revision"] = "Amazon Redshift datashare assets per revision";
159
- LimitName["Amazon_S3_data_access_assets_per_revision"] = "Amazon S3 data access assets per revision";
160
- LimitName["Asset_per_export_job_from_Amazon_S3"] = "Asset per export job from Amazon S3";
161
- LimitName["Asset_size_in_GB"] = "Asset size in GB";
162
- LimitName["Assets_per_import_job_from_Amazon_S3"] = "Assets per import job from Amazon S3";
163
- LimitName["Assets_per_revision"] = "Assets per revision";
164
- LimitName["Auto_export_event_actions_per_data_set"] = "Auto export event actions per data set";
165
- LimitName["Concurrent_in_progress_jobs_to_create_Amazon_S3_data_access_assets_from_S3_buckets"] = "Concurrent in progress jobs to create Amazon S3 data access assets from S3 buckets";
166
- LimitName["Concurrent_in_progress_jobs_to_export_assets_to_Amazon_S3"] = "Concurrent in progress jobs to export assets to Amazon S3";
167
- LimitName["Concurrent_in_progress_jobs_to_export_assets_to_a_signed_URL"] = "Concurrent in progress jobs to export assets to a signed URL";
168
- LimitName["Concurrent_in_progress_jobs_to_export_revisions_to_Amazon_S3"] = "Concurrent in progress jobs to export revisions to Amazon S3";
169
- LimitName["Concurrent_in_progress_jobs_to_import_assets_from_Amazon_Redshift_datashares"] = "Concurrent in progress jobs to import assets from Amazon Redshift datashares";
170
- LimitName["Concurrent_in_progress_jobs_to_import_assets_from_Amazon_S3"] = "Concurrent in progress jobs to import assets from Amazon S3";
171
- LimitName["Concurrent_in_progress_jobs_to_import_assets_from_a_signed_URL"] = "Concurrent in progress jobs to import assets from a signed URL";
172
- LimitName["Concurrent_in_progress_jobs_to_import_assets_from_an_API_Gateway_API"] = "Concurrent in progress jobs to import assets from an API Gateway API";
173
- LimitName["Concurrent_in_progress_jobs_to_import_assets_from_an_AWS_Lake_Formation_tag_policy"] = "Concurrent in progress jobs to import assets from an AWS Lake Formation tag policy";
174
- LimitName["Data_sets_per_account"] = "Data sets per account";
175
- LimitName["Data_sets_per_product"] = "Data sets per product";
176
- LimitName["Event_actions_per_account"] = "Event actions per account";
177
- LimitName["Products_per_account"] = "Products per account";
178
- LimitName["Revisions_per_AWS_Lake_Formation_data_permission_data_set"] = "Revisions per AWS Lake Formation data permission data set";
179
- LimitName["Revisions_per_Amazon_API_Gateway_API_data_set"] = "Revisions per Amazon API Gateway API data set";
180
- LimitName["Revisions_per_Amazon_Redshift_datashare_data_set"] = "Revisions per Amazon Redshift datashare data set";
181
- LimitName["Revisions_per_Amazon_S3_data_access_data_set"] = "Revisions per Amazon S3 data access data set";
182
- LimitName["Revisions_per_data_set"] = "Revisions per data set";
183
- })(LimitName = exports.LimitName || (exports.LimitName = {}));
130
+ exports.Code = {
131
+ ACCESS_DENIED_EXCEPTION: "ACCESS_DENIED_EXCEPTION",
132
+ INTERNAL_SERVER_EXCEPTION: "INTERNAL_SERVER_EXCEPTION",
133
+ MALWARE_DETECTED: "MALWARE_DETECTED",
134
+ MALWARE_SCAN_ENCRYPTED_FILE: "MALWARE_SCAN_ENCRYPTED_FILE",
135
+ RESOURCE_NOT_FOUND_EXCEPTION: "RESOURCE_NOT_FOUND_EXCEPTION",
136
+ SERVICE_QUOTA_EXCEEDED_EXCEPTION: "SERVICE_QUOTA_EXCEEDED_EXCEPTION",
137
+ VALIDATION_EXCEPTION: "VALIDATION_EXCEPTION",
138
+ };
139
+ exports.Origin = {
140
+ ENTITLED: "ENTITLED",
141
+ OWNED: "OWNED",
142
+ };
143
+ exports.LimitName = {
144
+ AWS_Lake_Formation_data_permission_assets_per_revision: "AWS Lake Formation data permission assets per revision",
145
+ Amazon_API_Gateway_API_assets_per_revision: "Amazon API Gateway API assets per revision",
146
+ Amazon_Redshift_datashare_assets_per_import_job_from_Redshift: "Amazon Redshift datashare assets per import job from Redshift",
147
+ Amazon_Redshift_datashare_assets_per_revision: "Amazon Redshift datashare assets per revision",
148
+ Amazon_S3_data_access_assets_per_revision: "Amazon S3 data access assets per revision",
149
+ Asset_per_export_job_from_Amazon_S3: "Asset per export job from Amazon S3",
150
+ Asset_size_in_GB: "Asset size in GB",
151
+ Assets_per_import_job_from_Amazon_S3: "Assets per import job from Amazon S3",
152
+ Assets_per_revision: "Assets per revision",
153
+ Auto_export_event_actions_per_data_set: "Auto export event actions per data set",
154
+ Concurrent_in_progress_jobs_to_create_Amazon_S3_data_access_assets_from_S3_buckets: "Concurrent in progress jobs to create Amazon S3 data access assets from S3 buckets",
155
+ Concurrent_in_progress_jobs_to_export_assets_to_Amazon_S3: "Concurrent in progress jobs to export assets to Amazon S3",
156
+ Concurrent_in_progress_jobs_to_export_assets_to_a_signed_URL: "Concurrent in progress jobs to export assets to a signed URL",
157
+ Concurrent_in_progress_jobs_to_export_revisions_to_Amazon_S3: "Concurrent in progress jobs to export revisions to Amazon S3",
158
+ Concurrent_in_progress_jobs_to_import_assets_from_Amazon_Redshift_datashares: "Concurrent in progress jobs to import assets from Amazon Redshift datashares",
159
+ Concurrent_in_progress_jobs_to_import_assets_from_Amazon_S3: "Concurrent in progress jobs to import assets from Amazon S3",
160
+ Concurrent_in_progress_jobs_to_import_assets_from_a_signed_URL: "Concurrent in progress jobs to import assets from a signed URL",
161
+ Concurrent_in_progress_jobs_to_import_assets_from_an_API_Gateway_API: "Concurrent in progress jobs to import assets from an API Gateway API",
162
+ Concurrent_in_progress_jobs_to_import_assets_from_an_AWS_Lake_Formation_tag_policy: "Concurrent in progress jobs to import assets from an AWS Lake Formation tag policy",
163
+ Data_sets_per_account: "Data sets per account",
164
+ Data_sets_per_product: "Data sets per product",
165
+ Event_actions_per_account: "Event actions per account",
166
+ Products_per_account: "Products per account",
167
+ Revisions_per_AWS_Lake_Formation_data_permission_data_set: "Revisions per AWS Lake Formation data permission data set",
168
+ Revisions_per_Amazon_API_Gateway_API_data_set: "Revisions per Amazon API Gateway API data set",
169
+ Revisions_per_Amazon_Redshift_datashare_data_set: "Revisions per Amazon Redshift datashare data set",
170
+ Revisions_per_Amazon_S3_data_access_data_set: "Revisions per Amazon S3 data access data set",
171
+ Revisions_per_data_set: "Revisions per data set",
172
+ };
184
173
  class ServiceLimitExceededException extends DataExchangeServiceException_1.DataExchangeServiceException {
185
174
  constructor(opts) {
186
175
  super({
@@ -197,47 +186,41 @@ class ServiceLimitExceededException extends DataExchangeServiceException_1.DataE
197
186
  }
198
187
  }
199
188
  exports.ServiceLimitExceededException = ServiceLimitExceededException;
200
- var DatabaseLFTagPolicyPermission;
201
- (function (DatabaseLFTagPolicyPermission) {
202
- DatabaseLFTagPolicyPermission["DESCRIBE"] = "DESCRIBE";
203
- })(DatabaseLFTagPolicyPermission = exports.DatabaseLFTagPolicyPermission || (exports.DatabaseLFTagPolicyPermission = {}));
204
- var TableTagPolicyLFPermission;
205
- (function (TableTagPolicyLFPermission) {
206
- TableTagPolicyLFPermission["DESCRIBE"] = "DESCRIBE";
207
- TableTagPolicyLFPermission["SELECT"] = "SELECT";
208
- })(TableTagPolicyLFPermission = exports.TableTagPolicyLFPermission || (exports.TableTagPolicyLFPermission = {}));
209
- var Type;
210
- (function (Type) {
211
- Type["CREATE_S3_DATA_ACCESS_FROM_S3_BUCKET"] = "CREATE_S3_DATA_ACCESS_FROM_S3_BUCKET";
212
- Type["EXPORT_ASSETS_TO_S3"] = "EXPORT_ASSETS_TO_S3";
213
- Type["EXPORT_ASSET_TO_SIGNED_URL"] = "EXPORT_ASSET_TO_SIGNED_URL";
214
- Type["EXPORT_REVISIONS_TO_S3"] = "EXPORT_REVISIONS_TO_S3";
215
- Type["IMPORT_ASSETS_FROM_LAKE_FORMATION_TAG_POLICY"] = "IMPORT_ASSETS_FROM_LAKE_FORMATION_TAG_POLICY";
216
- Type["IMPORT_ASSETS_FROM_REDSHIFT_DATA_SHARES"] = "IMPORT_ASSETS_FROM_REDSHIFT_DATA_SHARES";
217
- Type["IMPORT_ASSETS_FROM_S3"] = "IMPORT_ASSETS_FROM_S3";
218
- Type["IMPORT_ASSET_FROM_API_GATEWAY_API"] = "IMPORT_ASSET_FROM_API_GATEWAY_API";
219
- Type["IMPORT_ASSET_FROM_SIGNED_URL"] = "IMPORT_ASSET_FROM_SIGNED_URL";
220
- })(Type = exports.Type || (exports.Type = {}));
221
- var JobErrorLimitName;
222
- (function (JobErrorLimitName) {
223
- JobErrorLimitName["AWS_Lake_Formation_data_permission_assets_per_revision"] = "AWS Lake Formation data permission assets per revision";
224
- JobErrorLimitName["Amazon_Redshift_datashare_assets_per_revision"] = "Amazon Redshift datashare assets per revision";
225
- JobErrorLimitName["Amazon_S3_data_access_assets_per_revision"] = "Amazon S3 data access assets per revision";
226
- JobErrorLimitName["Asset_size_in_GB"] = "Asset size in GB";
227
- JobErrorLimitName["Assets_per_revision"] = "Assets per revision";
228
- })(JobErrorLimitName = exports.JobErrorLimitName || (exports.JobErrorLimitName = {}));
229
- var JobErrorResourceTypes;
230
- (function (JobErrorResourceTypes) {
231
- JobErrorResourceTypes["ASSET"] = "ASSET";
232
- JobErrorResourceTypes["DATA_SET"] = "DATA_SET";
233
- JobErrorResourceTypes["REVISION"] = "REVISION";
234
- })(JobErrorResourceTypes = exports.JobErrorResourceTypes || (exports.JobErrorResourceTypes = {}));
235
- var State;
236
- (function (State) {
237
- State["CANCELLED"] = "CANCELLED";
238
- State["COMPLETED"] = "COMPLETED";
239
- State["ERROR"] = "ERROR";
240
- State["IN_PROGRESS"] = "IN_PROGRESS";
241
- State["TIMED_OUT"] = "TIMED_OUT";
242
- State["WAITING"] = "WAITING";
243
- })(State = exports.State || (exports.State = {}));
189
+ exports.DatabaseLFTagPolicyPermission = {
190
+ DESCRIBE: "DESCRIBE",
191
+ };
192
+ exports.TableTagPolicyLFPermission = {
193
+ DESCRIBE: "DESCRIBE",
194
+ SELECT: "SELECT",
195
+ };
196
+ exports.Type = {
197
+ CREATE_S3_DATA_ACCESS_FROM_S3_BUCKET: "CREATE_S3_DATA_ACCESS_FROM_S3_BUCKET",
198
+ EXPORT_ASSETS_TO_S3: "EXPORT_ASSETS_TO_S3",
199
+ EXPORT_ASSET_TO_SIGNED_URL: "EXPORT_ASSET_TO_SIGNED_URL",
200
+ EXPORT_REVISIONS_TO_S3: "EXPORT_REVISIONS_TO_S3",
201
+ IMPORT_ASSETS_FROM_LAKE_FORMATION_TAG_POLICY: "IMPORT_ASSETS_FROM_LAKE_FORMATION_TAG_POLICY",
202
+ IMPORT_ASSETS_FROM_REDSHIFT_DATA_SHARES: "IMPORT_ASSETS_FROM_REDSHIFT_DATA_SHARES",
203
+ IMPORT_ASSETS_FROM_S3: "IMPORT_ASSETS_FROM_S3",
204
+ IMPORT_ASSET_FROM_API_GATEWAY_API: "IMPORT_ASSET_FROM_API_GATEWAY_API",
205
+ IMPORT_ASSET_FROM_SIGNED_URL: "IMPORT_ASSET_FROM_SIGNED_URL",
206
+ };
207
+ exports.JobErrorLimitName = {
208
+ AWS_Lake_Formation_data_permission_assets_per_revision: "AWS Lake Formation data permission assets per revision",
209
+ Amazon_Redshift_datashare_assets_per_revision: "Amazon Redshift datashare assets per revision",
210
+ Amazon_S3_data_access_assets_per_revision: "Amazon S3 data access assets per revision",
211
+ Asset_size_in_GB: "Asset size in GB",
212
+ Assets_per_revision: "Assets per revision",
213
+ };
214
+ exports.JobErrorResourceTypes = {
215
+ ASSET: "ASSET",
216
+ DATA_SET: "DATA_SET",
217
+ REVISION: "REVISION",
218
+ };
219
+ exports.State = {
220
+ CANCELLED: "CANCELLED",
221
+ COMPLETED: "COMPLETED",
222
+ ERROR: "ERROR",
223
+ IN_PROGRESS: "IN_PROGRESS",
224
+ TIMED_OUT: "TIMED_OUT",
225
+ WAITING: "WAITING",
226
+ };
@@ -12,45 +12,38 @@ export class AccessDeniedException extends __BaseException {
12
12
  this.Message = opts.Message;
13
13
  }
14
14
  }
15
- export var ServerSideEncryptionTypes;
16
- (function (ServerSideEncryptionTypes) {
17
- ServerSideEncryptionTypes["AES256"] = "AES256";
18
- ServerSideEncryptionTypes["aws_kms"] = "aws:kms";
19
- })(ServerSideEncryptionTypes || (ServerSideEncryptionTypes = {}));
20
- export var ProtocolType;
21
- (function (ProtocolType) {
22
- ProtocolType["REST"] = "REST";
23
- })(ProtocolType || (ProtocolType = {}));
24
- export var LFResourceType;
25
- (function (LFResourceType) {
26
- LFResourceType["DATABASE"] = "DATABASE";
27
- LFResourceType["TABLE"] = "TABLE";
28
- })(LFResourceType || (LFResourceType = {}));
29
- export var LakeFormationDataPermissionType;
30
- (function (LakeFormationDataPermissionType) {
31
- LakeFormationDataPermissionType["LFTagPolicy"] = "LFTagPolicy";
32
- })(LakeFormationDataPermissionType || (LakeFormationDataPermissionType = {}));
33
- export var LFPermission;
34
- (function (LFPermission) {
35
- LFPermission["DESCRIBE"] = "DESCRIBE";
36
- LFPermission["SELECT"] = "SELECT";
37
- })(LFPermission || (LFPermission = {}));
38
- export var AssetType;
39
- (function (AssetType) {
40
- AssetType["API_GATEWAY_API"] = "API_GATEWAY_API";
41
- AssetType["LAKE_FORMATION_DATA_PERMISSION"] = "LAKE_FORMATION_DATA_PERMISSION";
42
- AssetType["REDSHIFT_DATA_SHARE"] = "REDSHIFT_DATA_SHARE";
43
- AssetType["S3_DATA_ACCESS"] = "S3_DATA_ACCESS";
44
- AssetType["S3_SNAPSHOT"] = "S3_SNAPSHOT";
45
- })(AssetType || (AssetType = {}));
46
- export var ResourceType;
47
- (function (ResourceType) {
48
- ResourceType["ASSET"] = "ASSET";
49
- ResourceType["DATA_SET"] = "DATA_SET";
50
- ResourceType["EVENT_ACTION"] = "EVENT_ACTION";
51
- ResourceType["JOB"] = "JOB";
52
- ResourceType["REVISION"] = "REVISION";
53
- })(ResourceType || (ResourceType = {}));
15
+ export const ServerSideEncryptionTypes = {
16
+ AES256: "AES256",
17
+ aws_kms: "aws:kms",
18
+ };
19
+ export const ProtocolType = {
20
+ REST: "REST",
21
+ };
22
+ export const LFResourceType = {
23
+ DATABASE: "DATABASE",
24
+ TABLE: "TABLE",
25
+ };
26
+ export const LakeFormationDataPermissionType = {
27
+ LFTagPolicy: "LFTagPolicy",
28
+ };
29
+ export const LFPermission = {
30
+ DESCRIBE: "DESCRIBE",
31
+ SELECT: "SELECT",
32
+ };
33
+ export const AssetType = {
34
+ API_GATEWAY_API: "API_GATEWAY_API",
35
+ LAKE_FORMATION_DATA_PERMISSION: "LAKE_FORMATION_DATA_PERMISSION",
36
+ REDSHIFT_DATA_SHARE: "REDSHIFT_DATA_SHARE",
37
+ S3_DATA_ACCESS: "S3_DATA_ACCESS",
38
+ S3_SNAPSHOT: "S3_SNAPSHOT",
39
+ };
40
+ export const ResourceType = {
41
+ ASSET: "ASSET",
42
+ DATA_SET: "DATA_SET",
43
+ EVENT_ACTION: "EVENT_ACTION",
44
+ JOB: "JOB",
45
+ REVISION: "REVISION",
46
+ };
54
47
  export class ConflictException extends __BaseException {
55
48
  constructor(opts) {
56
49
  super({
@@ -107,11 +100,10 @@ export class ThrottlingException extends __BaseException {
107
100
  this.Message = opts.Message;
108
101
  }
109
102
  }
110
- export var ExceptionCause;
111
- (function (ExceptionCause) {
112
- ExceptionCause["InsufficientS3BucketPolicy"] = "InsufficientS3BucketPolicy";
113
- ExceptionCause["S3AccessDenied"] = "S3AccessDenied";
114
- })(ExceptionCause || (ExceptionCause = {}));
103
+ export const ExceptionCause = {
104
+ InsufficientS3BucketPolicy: "InsufficientS3BucketPolicy",
105
+ S3AccessDenied: "S3AccessDenied",
106
+ };
115
107
  export class ValidationException extends __BaseException {
116
108
  constructor(opts) {
117
109
  super({
@@ -126,52 +118,49 @@ export class ValidationException extends __BaseException {
126
118
  this.ExceptionCause = opts.ExceptionCause;
127
119
  }
128
120
  }
129
- export var Code;
130
- (function (Code) {
131
- Code["ACCESS_DENIED_EXCEPTION"] = "ACCESS_DENIED_EXCEPTION";
132
- Code["INTERNAL_SERVER_EXCEPTION"] = "INTERNAL_SERVER_EXCEPTION";
133
- Code["MALWARE_DETECTED"] = "MALWARE_DETECTED";
134
- Code["MALWARE_SCAN_ENCRYPTED_FILE"] = "MALWARE_SCAN_ENCRYPTED_FILE";
135
- Code["RESOURCE_NOT_FOUND_EXCEPTION"] = "RESOURCE_NOT_FOUND_EXCEPTION";
136
- Code["SERVICE_QUOTA_EXCEEDED_EXCEPTION"] = "SERVICE_QUOTA_EXCEEDED_EXCEPTION";
137
- Code["VALIDATION_EXCEPTION"] = "VALIDATION_EXCEPTION";
138
- })(Code || (Code = {}));
139
- export var Origin;
140
- (function (Origin) {
141
- Origin["ENTITLED"] = "ENTITLED";
142
- Origin["OWNED"] = "OWNED";
143
- })(Origin || (Origin = {}));
144
- export var LimitName;
145
- (function (LimitName) {
146
- LimitName["AWS_Lake_Formation_data_permission_assets_per_revision"] = "AWS Lake Formation data permission assets per revision";
147
- LimitName["Amazon_API_Gateway_API_assets_per_revision"] = "Amazon API Gateway API assets per revision";
148
- LimitName["Amazon_Redshift_datashare_assets_per_import_job_from_Redshift"] = "Amazon Redshift datashare assets per import job from Redshift";
149
- LimitName["Amazon_Redshift_datashare_assets_per_revision"] = "Amazon Redshift datashare assets per revision";
150
- LimitName["Amazon_S3_data_access_assets_per_revision"] = "Amazon S3 data access assets per revision";
151
- LimitName["Asset_per_export_job_from_Amazon_S3"] = "Asset per export job from Amazon S3";
152
- LimitName["Asset_size_in_GB"] = "Asset size in GB";
153
- LimitName["Assets_per_import_job_from_Amazon_S3"] = "Assets per import job from Amazon S3";
154
- LimitName["Assets_per_revision"] = "Assets per revision";
155
- LimitName["Auto_export_event_actions_per_data_set"] = "Auto export event actions per data set";
156
- LimitName["Concurrent_in_progress_jobs_to_create_Amazon_S3_data_access_assets_from_S3_buckets"] = "Concurrent in progress jobs to create Amazon S3 data access assets from S3 buckets";
157
- LimitName["Concurrent_in_progress_jobs_to_export_assets_to_Amazon_S3"] = "Concurrent in progress jobs to export assets to Amazon S3";
158
- LimitName["Concurrent_in_progress_jobs_to_export_assets_to_a_signed_URL"] = "Concurrent in progress jobs to export assets to a signed URL";
159
- LimitName["Concurrent_in_progress_jobs_to_export_revisions_to_Amazon_S3"] = "Concurrent in progress jobs to export revisions to Amazon S3";
160
- LimitName["Concurrent_in_progress_jobs_to_import_assets_from_Amazon_Redshift_datashares"] = "Concurrent in progress jobs to import assets from Amazon Redshift datashares";
161
- LimitName["Concurrent_in_progress_jobs_to_import_assets_from_Amazon_S3"] = "Concurrent in progress jobs to import assets from Amazon S3";
162
- LimitName["Concurrent_in_progress_jobs_to_import_assets_from_a_signed_URL"] = "Concurrent in progress jobs to import assets from a signed URL";
163
- LimitName["Concurrent_in_progress_jobs_to_import_assets_from_an_API_Gateway_API"] = "Concurrent in progress jobs to import assets from an API Gateway API";
164
- LimitName["Concurrent_in_progress_jobs_to_import_assets_from_an_AWS_Lake_Formation_tag_policy"] = "Concurrent in progress jobs to import assets from an AWS Lake Formation tag policy";
165
- LimitName["Data_sets_per_account"] = "Data sets per account";
166
- LimitName["Data_sets_per_product"] = "Data sets per product";
167
- LimitName["Event_actions_per_account"] = "Event actions per account";
168
- LimitName["Products_per_account"] = "Products per account";
169
- LimitName["Revisions_per_AWS_Lake_Formation_data_permission_data_set"] = "Revisions per AWS Lake Formation data permission data set";
170
- LimitName["Revisions_per_Amazon_API_Gateway_API_data_set"] = "Revisions per Amazon API Gateway API data set";
171
- LimitName["Revisions_per_Amazon_Redshift_datashare_data_set"] = "Revisions per Amazon Redshift datashare data set";
172
- LimitName["Revisions_per_Amazon_S3_data_access_data_set"] = "Revisions per Amazon S3 data access data set";
173
- LimitName["Revisions_per_data_set"] = "Revisions per data set";
174
- })(LimitName || (LimitName = {}));
121
+ export const Code = {
122
+ ACCESS_DENIED_EXCEPTION: "ACCESS_DENIED_EXCEPTION",
123
+ INTERNAL_SERVER_EXCEPTION: "INTERNAL_SERVER_EXCEPTION",
124
+ MALWARE_DETECTED: "MALWARE_DETECTED",
125
+ MALWARE_SCAN_ENCRYPTED_FILE: "MALWARE_SCAN_ENCRYPTED_FILE",
126
+ RESOURCE_NOT_FOUND_EXCEPTION: "RESOURCE_NOT_FOUND_EXCEPTION",
127
+ SERVICE_QUOTA_EXCEEDED_EXCEPTION: "SERVICE_QUOTA_EXCEEDED_EXCEPTION",
128
+ VALIDATION_EXCEPTION: "VALIDATION_EXCEPTION",
129
+ };
130
+ export const Origin = {
131
+ ENTITLED: "ENTITLED",
132
+ OWNED: "OWNED",
133
+ };
134
+ export const LimitName = {
135
+ AWS_Lake_Formation_data_permission_assets_per_revision: "AWS Lake Formation data permission assets per revision",
136
+ Amazon_API_Gateway_API_assets_per_revision: "Amazon API Gateway API assets per revision",
137
+ Amazon_Redshift_datashare_assets_per_import_job_from_Redshift: "Amazon Redshift datashare assets per import job from Redshift",
138
+ Amazon_Redshift_datashare_assets_per_revision: "Amazon Redshift datashare assets per revision",
139
+ Amazon_S3_data_access_assets_per_revision: "Amazon S3 data access assets per revision",
140
+ Asset_per_export_job_from_Amazon_S3: "Asset per export job from Amazon S3",
141
+ Asset_size_in_GB: "Asset size in GB",
142
+ Assets_per_import_job_from_Amazon_S3: "Assets per import job from Amazon S3",
143
+ Assets_per_revision: "Assets per revision",
144
+ Auto_export_event_actions_per_data_set: "Auto export event actions per data set",
145
+ Concurrent_in_progress_jobs_to_create_Amazon_S3_data_access_assets_from_S3_buckets: "Concurrent in progress jobs to create Amazon S3 data access assets from S3 buckets",
146
+ Concurrent_in_progress_jobs_to_export_assets_to_Amazon_S3: "Concurrent in progress jobs to export assets to Amazon S3",
147
+ Concurrent_in_progress_jobs_to_export_assets_to_a_signed_URL: "Concurrent in progress jobs to export assets to a signed URL",
148
+ Concurrent_in_progress_jobs_to_export_revisions_to_Amazon_S3: "Concurrent in progress jobs to export revisions to Amazon S3",
149
+ Concurrent_in_progress_jobs_to_import_assets_from_Amazon_Redshift_datashares: "Concurrent in progress jobs to import assets from Amazon Redshift datashares",
150
+ Concurrent_in_progress_jobs_to_import_assets_from_Amazon_S3: "Concurrent in progress jobs to import assets from Amazon S3",
151
+ Concurrent_in_progress_jobs_to_import_assets_from_a_signed_URL: "Concurrent in progress jobs to import assets from a signed URL",
152
+ Concurrent_in_progress_jobs_to_import_assets_from_an_API_Gateway_API: "Concurrent in progress jobs to import assets from an API Gateway API",
153
+ Concurrent_in_progress_jobs_to_import_assets_from_an_AWS_Lake_Formation_tag_policy: "Concurrent in progress jobs to import assets from an AWS Lake Formation tag policy",
154
+ Data_sets_per_account: "Data sets per account",
155
+ Data_sets_per_product: "Data sets per product",
156
+ Event_actions_per_account: "Event actions per account",
157
+ Products_per_account: "Products per account",
158
+ Revisions_per_AWS_Lake_Formation_data_permission_data_set: "Revisions per AWS Lake Formation data permission data set",
159
+ Revisions_per_Amazon_API_Gateway_API_data_set: "Revisions per Amazon API Gateway API data set",
160
+ Revisions_per_Amazon_Redshift_datashare_data_set: "Revisions per Amazon Redshift datashare data set",
161
+ Revisions_per_Amazon_S3_data_access_data_set: "Revisions per Amazon S3 data access data set",
162
+ Revisions_per_data_set: "Revisions per data set",
163
+ };
175
164
  export class ServiceLimitExceededException extends __BaseException {
176
165
  constructor(opts) {
177
166
  super({
@@ -187,47 +176,41 @@ export class ServiceLimitExceededException extends __BaseException {
187
176
  this.Message = opts.Message;
188
177
  }
189
178
  }
190
- export var DatabaseLFTagPolicyPermission;
191
- (function (DatabaseLFTagPolicyPermission) {
192
- DatabaseLFTagPolicyPermission["DESCRIBE"] = "DESCRIBE";
193
- })(DatabaseLFTagPolicyPermission || (DatabaseLFTagPolicyPermission = {}));
194
- export var TableTagPolicyLFPermission;
195
- (function (TableTagPolicyLFPermission) {
196
- TableTagPolicyLFPermission["DESCRIBE"] = "DESCRIBE";
197
- TableTagPolicyLFPermission["SELECT"] = "SELECT";
198
- })(TableTagPolicyLFPermission || (TableTagPolicyLFPermission = {}));
199
- export var Type;
200
- (function (Type) {
201
- Type["CREATE_S3_DATA_ACCESS_FROM_S3_BUCKET"] = "CREATE_S3_DATA_ACCESS_FROM_S3_BUCKET";
202
- Type["EXPORT_ASSETS_TO_S3"] = "EXPORT_ASSETS_TO_S3";
203
- Type["EXPORT_ASSET_TO_SIGNED_URL"] = "EXPORT_ASSET_TO_SIGNED_URL";
204
- Type["EXPORT_REVISIONS_TO_S3"] = "EXPORT_REVISIONS_TO_S3";
205
- Type["IMPORT_ASSETS_FROM_LAKE_FORMATION_TAG_POLICY"] = "IMPORT_ASSETS_FROM_LAKE_FORMATION_TAG_POLICY";
206
- Type["IMPORT_ASSETS_FROM_REDSHIFT_DATA_SHARES"] = "IMPORT_ASSETS_FROM_REDSHIFT_DATA_SHARES";
207
- Type["IMPORT_ASSETS_FROM_S3"] = "IMPORT_ASSETS_FROM_S3";
208
- Type["IMPORT_ASSET_FROM_API_GATEWAY_API"] = "IMPORT_ASSET_FROM_API_GATEWAY_API";
209
- Type["IMPORT_ASSET_FROM_SIGNED_URL"] = "IMPORT_ASSET_FROM_SIGNED_URL";
210
- })(Type || (Type = {}));
211
- export var JobErrorLimitName;
212
- (function (JobErrorLimitName) {
213
- JobErrorLimitName["AWS_Lake_Formation_data_permission_assets_per_revision"] = "AWS Lake Formation data permission assets per revision";
214
- JobErrorLimitName["Amazon_Redshift_datashare_assets_per_revision"] = "Amazon Redshift datashare assets per revision";
215
- JobErrorLimitName["Amazon_S3_data_access_assets_per_revision"] = "Amazon S3 data access assets per revision";
216
- JobErrorLimitName["Asset_size_in_GB"] = "Asset size in GB";
217
- JobErrorLimitName["Assets_per_revision"] = "Assets per revision";
218
- })(JobErrorLimitName || (JobErrorLimitName = {}));
219
- export var JobErrorResourceTypes;
220
- (function (JobErrorResourceTypes) {
221
- JobErrorResourceTypes["ASSET"] = "ASSET";
222
- JobErrorResourceTypes["DATA_SET"] = "DATA_SET";
223
- JobErrorResourceTypes["REVISION"] = "REVISION";
224
- })(JobErrorResourceTypes || (JobErrorResourceTypes = {}));
225
- export var State;
226
- (function (State) {
227
- State["CANCELLED"] = "CANCELLED";
228
- State["COMPLETED"] = "COMPLETED";
229
- State["ERROR"] = "ERROR";
230
- State["IN_PROGRESS"] = "IN_PROGRESS";
231
- State["TIMED_OUT"] = "TIMED_OUT";
232
- State["WAITING"] = "WAITING";
233
- })(State || (State = {}));
179
+ export const DatabaseLFTagPolicyPermission = {
180
+ DESCRIBE: "DESCRIBE",
181
+ };
182
+ export const TableTagPolicyLFPermission = {
183
+ DESCRIBE: "DESCRIBE",
184
+ SELECT: "SELECT",
185
+ };
186
+ export const Type = {
187
+ CREATE_S3_DATA_ACCESS_FROM_S3_BUCKET: "CREATE_S3_DATA_ACCESS_FROM_S3_BUCKET",
188
+ EXPORT_ASSETS_TO_S3: "EXPORT_ASSETS_TO_S3",
189
+ EXPORT_ASSET_TO_SIGNED_URL: "EXPORT_ASSET_TO_SIGNED_URL",
190
+ EXPORT_REVISIONS_TO_S3: "EXPORT_REVISIONS_TO_S3",
191
+ IMPORT_ASSETS_FROM_LAKE_FORMATION_TAG_POLICY: "IMPORT_ASSETS_FROM_LAKE_FORMATION_TAG_POLICY",
192
+ IMPORT_ASSETS_FROM_REDSHIFT_DATA_SHARES: "IMPORT_ASSETS_FROM_REDSHIFT_DATA_SHARES",
193
+ IMPORT_ASSETS_FROM_S3: "IMPORT_ASSETS_FROM_S3",
194
+ IMPORT_ASSET_FROM_API_GATEWAY_API: "IMPORT_ASSET_FROM_API_GATEWAY_API",
195
+ IMPORT_ASSET_FROM_SIGNED_URL: "IMPORT_ASSET_FROM_SIGNED_URL",
196
+ };
197
+ export const JobErrorLimitName = {
198
+ AWS_Lake_Formation_data_permission_assets_per_revision: "AWS Lake Formation data permission assets per revision",
199
+ Amazon_Redshift_datashare_assets_per_revision: "Amazon Redshift datashare assets per revision",
200
+ Amazon_S3_data_access_assets_per_revision: "Amazon S3 data access assets per revision",
201
+ Asset_size_in_GB: "Asset size in GB",
202
+ Assets_per_revision: "Assets per revision",
203
+ };
204
+ export const JobErrorResourceTypes = {
205
+ ASSET: "ASSET",
206
+ DATA_SET: "DATA_SET",
207
+ REVISION: "REVISION",
208
+ };
209
+ export const State = {
210
+ CANCELLED: "CANCELLED",
211
+ COMPLETED: "COMPLETED",
212
+ ERROR: "ERROR",
213
+ IN_PROGRESS: "IN_PROGRESS",
214
+ TIMED_OUT: "TIMED_OUT",
215
+ WAITING: "WAITING",
216
+ };