cdk-lambda-subminute 2.0.475 → 2.0.477
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.jsii +3 -3
- package/lib/cdk-lambda-subminute.js +3 -3
- package/node_modules/aws-sdk/Crashpad/settings.dat +0 -0
- package/node_modules/aws-sdk/README.md +1 -1
- package/node_modules/aws-sdk/apis/bedrock-2023-04-20.min.json +295 -29
- package/node_modules/aws-sdk/apis/bedrock-2023-04-20.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/codebuild-2016-10-06.min.json +3 -0
- package/node_modules/aws-sdk/apis/deadline-2023-10-12.min.json +1679 -1672
- package/node_modules/aws-sdk/apis/lambda-2015-03-31.min.json +81 -23
- package/node_modules/aws-sdk/apis/opensearchserverless-2021-11-01.min.json +11 -1
- package/node_modules/aws-sdk/apis/s3-2006-03-01.examples.json +79 -79
- package/node_modules/aws-sdk/apis/s3-2006-03-01.min.json +188 -180
- package/node_modules/aws-sdk/apis/ssm-sap-2018-05-10.min.json +37 -24
- package/node_modules/aws-sdk/clients/bedrock.d.ts +288 -0
- package/node_modules/aws-sdk/clients/codebuild.d.ts +14 -2
- package/node_modules/aws-sdk/clients/deadline.d.ts +1562 -1562
- package/node_modules/aws-sdk/clients/ecs.d.ts +54 -54
- package/node_modules/aws-sdk/clients/lambda.d.ts +56 -10
- package/node_modules/aws-sdk/clients/opensearchserverless.d.ts +16 -0
- package/node_modules/aws-sdk/clients/s3.d.ts +25 -17
- package/node_modules/aws-sdk/clients/ssmsap.d.ts +10 -0
- package/node_modules/aws-sdk/dist/aws-sdk-core-react-native.js +1 -1
- package/node_modules/aws-sdk/dist/aws-sdk-react-native.js +9 -9
- package/node_modules/aws-sdk/dist/aws-sdk.js +275 -206
- package/node_modules/aws-sdk/dist/aws-sdk.min.js +89 -89
- package/node_modules/aws-sdk/lib/core.js +1 -1
- package/node_modules/aws-sdk/package.json +1 -1
- package/package.json +4 -4
@@ -2,9 +2,14 @@
|
|
2
2
|
"version": "2.0",
|
3
3
|
"metadata": {
|
4
4
|
"apiVersion": "2018-05-10",
|
5
|
+
"auth": [
|
6
|
+
"aws.auth#sigv4"
|
7
|
+
],
|
5
8
|
"endpointPrefix": "ssm-sap",
|
6
|
-
"jsonVersion": "1.1",
|
7
9
|
"protocol": "rest-json",
|
10
|
+
"protocols": [
|
11
|
+
"rest-json"
|
12
|
+
],
|
8
13
|
"serviceAbbreviation": "SsmSap",
|
9
14
|
"serviceFullName": "AWS Systems Manager for SAP",
|
10
15
|
"serviceId": "Ssm Sap",
|
@@ -75,7 +80,7 @@
|
|
75
80
|
"shape": "Sd"
|
76
81
|
},
|
77
82
|
"Tags": {
|
78
|
-
"shape": "
|
83
|
+
"shape": "Sl"
|
79
84
|
}
|
80
85
|
}
|
81
86
|
}
|
@@ -188,7 +193,7 @@
|
|
188
193
|
}
|
189
194
|
},
|
190
195
|
"Tags": {
|
191
|
-
"shape": "
|
196
|
+
"shape": "Sl"
|
192
197
|
}
|
193
198
|
}
|
194
199
|
}
|
@@ -216,7 +221,7 @@
|
|
216
221
|
"ApplicationId": {},
|
217
222
|
"ComponentId": {},
|
218
223
|
"Credentials": {
|
219
|
-
"shape": "
|
224
|
+
"shape": "S1e"
|
220
225
|
},
|
221
226
|
"DatabaseId": {},
|
222
227
|
"DatabaseName": {},
|
@@ -229,11 +234,15 @@
|
|
229
234
|
},
|
230
235
|
"LastUpdated": {
|
231
236
|
"type": "timestamp"
|
237
|
+
},
|
238
|
+
"ConnectedComponentArns": {
|
239
|
+
"type": "list",
|
240
|
+
"member": {}
|
232
241
|
}
|
233
242
|
}
|
234
243
|
},
|
235
244
|
"Tags": {
|
236
|
-
"shape": "
|
245
|
+
"shape": "Sl"
|
237
246
|
}
|
238
247
|
}
|
239
248
|
}
|
@@ -256,7 +265,7 @@
|
|
256
265
|
"type": "structure",
|
257
266
|
"members": {
|
258
267
|
"Operation": {
|
259
|
-
"shape": "
|
268
|
+
"shape": "S1q"
|
260
269
|
}
|
261
270
|
}
|
262
271
|
}
|
@@ -296,7 +305,7 @@
|
|
296
305
|
"type": "integer"
|
297
306
|
},
|
298
307
|
"Filters": {
|
299
|
-
"shape": "
|
308
|
+
"shape": "S21"
|
300
309
|
}
|
301
310
|
}
|
302
311
|
},
|
@@ -313,7 +322,7 @@
|
|
313
322
|
"Type": {},
|
314
323
|
"Arn": {},
|
315
324
|
"Tags": {
|
316
|
-
"shape": "
|
325
|
+
"shape": "Sl"
|
317
326
|
}
|
318
327
|
}
|
319
328
|
}
|
@@ -349,7 +358,7 @@
|
|
349
358
|
"ComponentId": {},
|
350
359
|
"ComponentType": {},
|
351
360
|
"Tags": {
|
352
|
-
"shape": "
|
361
|
+
"shape": "Sl"
|
353
362
|
},
|
354
363
|
"Arn": {}
|
355
364
|
}
|
@@ -389,7 +398,7 @@
|
|
389
398
|
"DatabaseType": {},
|
390
399
|
"Arn": {},
|
391
400
|
"Tags": {
|
392
|
-
"shape": "
|
401
|
+
"shape": "Sl"
|
393
402
|
}
|
394
403
|
}
|
395
404
|
}
|
@@ -415,7 +424,7 @@
|
|
415
424
|
},
|
416
425
|
"NextToken": {},
|
417
426
|
"Filters": {
|
418
|
-
"shape": "
|
427
|
+
"shape": "S21"
|
419
428
|
}
|
420
429
|
}
|
421
430
|
},
|
@@ -464,7 +473,7 @@
|
|
464
473
|
},
|
465
474
|
"NextToken": {},
|
466
475
|
"Filters": {
|
467
|
-
"shape": "
|
476
|
+
"shape": "S21"
|
468
477
|
}
|
469
478
|
}
|
470
479
|
},
|
@@ -474,7 +483,7 @@
|
|
474
483
|
"Operations": {
|
475
484
|
"type": "list",
|
476
485
|
"member": {
|
477
|
-
"shape": "
|
486
|
+
"shape": "S1q"
|
478
487
|
}
|
479
488
|
},
|
480
489
|
"NextToken": {}
|
@@ -503,7 +512,7 @@
|
|
503
512
|
"type": "structure",
|
504
513
|
"members": {
|
505
514
|
"tags": {
|
506
|
-
"shape": "
|
515
|
+
"shape": "Sl"
|
507
516
|
}
|
508
517
|
}
|
509
518
|
}
|
@@ -555,10 +564,10 @@
|
|
555
564
|
"SapInstanceNumber": {},
|
556
565
|
"Sid": {},
|
557
566
|
"Tags": {
|
558
|
-
"shape": "
|
567
|
+
"shape": "Sl"
|
559
568
|
},
|
560
569
|
"Credentials": {
|
561
|
-
"shape": "
|
570
|
+
"shape": "S1e"
|
562
571
|
},
|
563
572
|
"DatabaseArn": {}
|
564
573
|
}
|
@@ -657,7 +666,7 @@
|
|
657
666
|
"locationName": "resourceArn"
|
658
667
|
},
|
659
668
|
"tags": {
|
660
|
-
"shape": "
|
669
|
+
"shape": "Sl"
|
661
670
|
}
|
662
671
|
}
|
663
672
|
},
|
@@ -710,10 +719,10 @@
|
|
710
719
|
"members": {
|
711
720
|
"ApplicationId": {},
|
712
721
|
"CredentialsToAddOrUpdate": {
|
713
|
-
"shape": "
|
722
|
+
"shape": "S1e"
|
714
723
|
},
|
715
724
|
"CredentialsToRemove": {
|
716
|
-
"shape": "
|
725
|
+
"shape": "S1e"
|
717
726
|
},
|
718
727
|
"Backint": {
|
719
728
|
"type": "structure",
|
@@ -759,19 +768,23 @@
|
|
759
768
|
"LastUpdated": {
|
760
769
|
"type": "timestamp"
|
761
770
|
},
|
762
|
-
"StatusMessage": {}
|
771
|
+
"StatusMessage": {},
|
772
|
+
"AssociatedApplicationArns": {
|
773
|
+
"type": "list",
|
774
|
+
"member": {}
|
775
|
+
}
|
763
776
|
}
|
764
777
|
},
|
765
778
|
"Sh": {
|
766
779
|
"type": "list",
|
767
780
|
"member": {}
|
768
781
|
},
|
769
|
-
"
|
782
|
+
"Sl": {
|
770
783
|
"type": "map",
|
771
784
|
"key": {},
|
772
785
|
"value": {}
|
773
786
|
},
|
774
|
-
"
|
787
|
+
"S1e": {
|
775
788
|
"type": "list",
|
776
789
|
"member": {
|
777
790
|
"type": "structure",
|
@@ -790,7 +803,7 @@
|
|
790
803
|
}
|
791
804
|
}
|
792
805
|
},
|
793
|
-
"
|
806
|
+
"S1q": {
|
794
807
|
"type": "structure",
|
795
808
|
"members": {
|
796
809
|
"Id": {},
|
@@ -816,7 +829,7 @@
|
|
816
829
|
}
|
817
830
|
}
|
818
831
|
},
|
819
|
-
"
|
832
|
+
"S21": {
|
820
833
|
"type": "list",
|
821
834
|
"member": {
|
822
835
|
"type": "structure",
|
@@ -51,6 +51,14 @@ declare class Bedrock extends Service {
|
|
51
51
|
* Creates a fine-tuning job to customize a base model. You specify the base foundation model and the location of the training data. After the model-customization job completes successfully, your custom model resource will be ready to use. Amazon Bedrock returns validation loss metrics and output generations after the job completes. For information on the format of training and validation data, see Prepare the datasets. Model-customization jobs are asynchronous and the completion time depends on the base model and the training/validation data size. To monitor a job, use the GetModelCustomizationJob operation to retrieve the job status. For more information, see Custom models in the Amazon Bedrock User Guide.
|
52
52
|
*/
|
53
53
|
createModelCustomizationJob(callback?: (err: AWSError, data: Bedrock.Types.CreateModelCustomizationJobResponse) => void): Request<Bedrock.Types.CreateModelCustomizationJobResponse, AWSError>;
|
54
|
+
/**
|
55
|
+
* Creates a job to invoke a model on multiple prompts (batch inference). Format your data according to Format your inference data and upload it to an Amazon S3 bucket. For more information, see Create a batch inference job. The response returns a jobArn that you can use to stop or get details about the job. You can check the status of the job by sending a GetModelCustomizationJob request.
|
56
|
+
*/
|
57
|
+
createModelInvocationJob(params: Bedrock.Types.CreateModelInvocationJobRequest, callback?: (err: AWSError, data: Bedrock.Types.CreateModelInvocationJobResponse) => void): Request<Bedrock.Types.CreateModelInvocationJobResponse, AWSError>;
|
58
|
+
/**
|
59
|
+
* Creates a job to invoke a model on multiple prompts (batch inference). Format your data according to Format your inference data and upload it to an Amazon S3 bucket. For more information, see Create a batch inference job. The response returns a jobArn that you can use to stop or get details about the job. You can check the status of the job by sending a GetModelCustomizationJob request.
|
60
|
+
*/
|
61
|
+
createModelInvocationJob(callback?: (err: AWSError, data: Bedrock.Types.CreateModelInvocationJobResponse) => void): Request<Bedrock.Types.CreateModelInvocationJobResponse, AWSError>;
|
54
62
|
/**
|
55
63
|
* Creates dedicated throughput for a base or custom model with the model units and for the duration that you specify. For pricing details, see Amazon Bedrock Pricing. For more information, see Provisioned Throughput in the Amazon Bedrock User Guide.
|
56
64
|
*/
|
@@ -139,6 +147,14 @@ declare class Bedrock extends Service {
|
|
139
147
|
* Retrieves the properties associated with a model-customization job, including the status of the job. For more information, see Custom models in the Amazon Bedrock User Guide.
|
140
148
|
*/
|
141
149
|
getModelCustomizationJob(callback?: (err: AWSError, data: Bedrock.Types.GetModelCustomizationJobResponse) => void): Request<Bedrock.Types.GetModelCustomizationJobResponse, AWSError>;
|
150
|
+
/**
|
151
|
+
* Gets details about a batch inference job. For more information, see View details about a batch inference job
|
152
|
+
*/
|
153
|
+
getModelInvocationJob(params: Bedrock.Types.GetModelInvocationJobRequest, callback?: (err: AWSError, data: Bedrock.Types.GetModelInvocationJobResponse) => void): Request<Bedrock.Types.GetModelInvocationJobResponse, AWSError>;
|
154
|
+
/**
|
155
|
+
* Gets details about a batch inference job. For more information, see View details about a batch inference job
|
156
|
+
*/
|
157
|
+
getModelInvocationJob(callback?: (err: AWSError, data: Bedrock.Types.GetModelInvocationJobResponse) => void): Request<Bedrock.Types.GetModelInvocationJobResponse, AWSError>;
|
142
158
|
/**
|
143
159
|
* Get the current configuration values for model invocation logging.
|
144
160
|
*/
|
@@ -203,6 +219,14 @@ declare class Bedrock extends Service {
|
|
203
219
|
* Returns a list of model customization jobs that you have submitted. You can filter the jobs to return based on one or more criteria. For more information, see Custom models in the Amazon Bedrock User Guide.
|
204
220
|
*/
|
205
221
|
listModelCustomizationJobs(callback?: (err: AWSError, data: Bedrock.Types.ListModelCustomizationJobsResponse) => void): Request<Bedrock.Types.ListModelCustomizationJobsResponse, AWSError>;
|
222
|
+
/**
|
223
|
+
* Lists all batch inference jobs in the account. For more information, see View details about a batch inference job.
|
224
|
+
*/
|
225
|
+
listModelInvocationJobs(params: Bedrock.Types.ListModelInvocationJobsRequest, callback?: (err: AWSError, data: Bedrock.Types.ListModelInvocationJobsResponse) => void): Request<Bedrock.Types.ListModelInvocationJobsResponse, AWSError>;
|
226
|
+
/**
|
227
|
+
* Lists all batch inference jobs in the account. For more information, see View details about a batch inference job.
|
228
|
+
*/
|
229
|
+
listModelInvocationJobs(callback?: (err: AWSError, data: Bedrock.Types.ListModelInvocationJobsResponse) => void): Request<Bedrock.Types.ListModelInvocationJobsResponse, AWSError>;
|
206
230
|
/**
|
207
231
|
* Lists the Provisioned Throughputs in the account. For more information, see Provisioned Throughput in the Amazon Bedrock User Guide.
|
208
232
|
*/
|
@@ -243,6 +267,14 @@ declare class Bedrock extends Service {
|
|
243
267
|
* Stops an active model customization job. For more information, see Custom models in the Amazon Bedrock User Guide.
|
244
268
|
*/
|
245
269
|
stopModelCustomizationJob(callback?: (err: AWSError, data: Bedrock.Types.StopModelCustomizationJobResponse) => void): Request<Bedrock.Types.StopModelCustomizationJobResponse, AWSError>;
|
270
|
+
/**
|
271
|
+
* Stops a batch inference job. You're only charged for tokens that were already processed. For more information, see Stop a batch inference job.
|
272
|
+
*/
|
273
|
+
stopModelInvocationJob(params: Bedrock.Types.StopModelInvocationJobRequest, callback?: (err: AWSError, data: Bedrock.Types.StopModelInvocationJobResponse) => void): Request<Bedrock.Types.StopModelInvocationJobResponse, AWSError>;
|
274
|
+
/**
|
275
|
+
* Stops a batch inference job. You're only charged for tokens that were already processed. For more information, see Stop a batch inference job.
|
276
|
+
*/
|
277
|
+
stopModelInvocationJob(callback?: (err: AWSError, data: Bedrock.Types.StopModelInvocationJobResponse) => void): Request<Bedrock.Types.StopModelInvocationJobResponse, AWSError>;
|
246
278
|
/**
|
247
279
|
* Associate tags with a resource. For more information, see Tagging resources in the Amazon Bedrock User Guide.
|
248
280
|
*/
|
@@ -532,6 +564,46 @@ declare namespace Bedrock {
|
|
532
564
|
*/
|
533
565
|
jobArn: ModelCustomizationJobArn;
|
534
566
|
}
|
567
|
+
export interface CreateModelInvocationJobRequest {
|
568
|
+
/**
|
569
|
+
* A name to give the batch inference job.
|
570
|
+
*/
|
571
|
+
jobName: ModelInvocationJobName;
|
572
|
+
/**
|
573
|
+
* The Amazon Resource Name (ARN) of the service role with permissions to carry out and manage batch inference. You can use the console to create a default service role or follow the steps at Create a service role for batch inference.
|
574
|
+
*/
|
575
|
+
roleArn: RoleArn;
|
576
|
+
/**
|
577
|
+
* A unique, case-sensitive identifier to ensure that the API request completes no more than one time. If this token matches a previous request, Amazon Bedrock ignores the request, but does not return an error. For more information, see Ensuring idempotency.
|
578
|
+
*/
|
579
|
+
clientRequestToken?: ModelInvocationIdempotencyToken;
|
580
|
+
/**
|
581
|
+
* The unique identifier of the foundation model to use for the batch inference job.
|
582
|
+
*/
|
583
|
+
modelId: ModelId;
|
584
|
+
/**
|
585
|
+
* Details about the location of the input to the batch inference job.
|
586
|
+
*/
|
587
|
+
inputDataConfig: ModelInvocationJobInputDataConfig;
|
588
|
+
/**
|
589
|
+
* Details about the location of the output of the batch inference job.
|
590
|
+
*/
|
591
|
+
outputDataConfig: ModelInvocationJobOutputDataConfig;
|
592
|
+
/**
|
593
|
+
* The number of hours after which to force the batch inference job to time out.
|
594
|
+
*/
|
595
|
+
timeoutDurationInHours?: ModelInvocationJobTimeoutDurationInHours;
|
596
|
+
/**
|
597
|
+
* Any tags to associate with the batch inference job. For more information, see Tagging Amazon Bedrock resources.
|
598
|
+
*/
|
599
|
+
tags?: TagList;
|
600
|
+
}
|
601
|
+
export interface CreateModelInvocationJobResponse {
|
602
|
+
/**
|
603
|
+
* The Amazon Resource Name (ARN) of the batch inference job.
|
604
|
+
*/
|
605
|
+
jobArn: ModelInvocationJobArn;
|
606
|
+
}
|
535
607
|
export interface CreateProvisionedModelThroughputRequest {
|
536
608
|
/**
|
537
609
|
* A unique, case-sensitive identifier to ensure that the API request completes no more than one time. If this token matches a previous request, Amazon Bedrock ignores the request, but does not return an error. For more information, see Ensuring idempotency in the Amazon S3 User Guide.
|
@@ -1207,6 +1279,70 @@ declare namespace Bedrock {
|
|
1207
1279
|
*/
|
1208
1280
|
vpcConfig?: VpcConfig;
|
1209
1281
|
}
|
1282
|
+
export interface GetModelInvocationJobRequest {
|
1283
|
+
/**
|
1284
|
+
* The Amazon Resource Name (ARN) of the batch inference job.
|
1285
|
+
*/
|
1286
|
+
jobIdentifier: ModelInvocationJobIdentifier;
|
1287
|
+
}
|
1288
|
+
export interface GetModelInvocationJobResponse {
|
1289
|
+
/**
|
1290
|
+
* The Amazon Resource Name (ARN) of the batch inference job.
|
1291
|
+
*/
|
1292
|
+
jobArn: ModelInvocationJobArn;
|
1293
|
+
/**
|
1294
|
+
* The name of the batch inference job.
|
1295
|
+
*/
|
1296
|
+
jobName?: ModelInvocationJobName;
|
1297
|
+
/**
|
1298
|
+
* The unique identifier of the foundation model used for model inference.
|
1299
|
+
*/
|
1300
|
+
modelId: ModelId;
|
1301
|
+
/**
|
1302
|
+
* A unique, case-sensitive identifier to ensure that the API request completes no more than one time. If this token matches a previous request, Amazon Bedrock ignores the request, but does not return an error. For more information, see Ensuring idempotency.
|
1303
|
+
*/
|
1304
|
+
clientRequestToken?: ModelInvocationIdempotencyToken;
|
1305
|
+
/**
|
1306
|
+
* The Amazon Resource Name (ARN) of the service role with permissions to carry out and manage batch inference. You can use the console to create a default service role or follow the steps at Create a service role for batch inference.
|
1307
|
+
*/
|
1308
|
+
roleArn: RoleArn;
|
1309
|
+
/**
|
1310
|
+
* The status of the batch inference job.
|
1311
|
+
*/
|
1312
|
+
status?: ModelInvocationJobStatus;
|
1313
|
+
/**
|
1314
|
+
* If the batch inference job failed, this field contains a message describing why the job failed.
|
1315
|
+
*/
|
1316
|
+
message?: Message;
|
1317
|
+
/**
|
1318
|
+
* The time at which the batch inference job was submitted.
|
1319
|
+
*/
|
1320
|
+
submitTime: Timestamp;
|
1321
|
+
/**
|
1322
|
+
* The time at which the batch inference job was last modified.
|
1323
|
+
*/
|
1324
|
+
lastModifiedTime?: Timestamp;
|
1325
|
+
/**
|
1326
|
+
* The time at which the batch inference job ended.
|
1327
|
+
*/
|
1328
|
+
endTime?: Timestamp;
|
1329
|
+
/**
|
1330
|
+
* Details about the location of the input to the batch inference job.
|
1331
|
+
*/
|
1332
|
+
inputDataConfig: ModelInvocationJobInputDataConfig;
|
1333
|
+
/**
|
1334
|
+
* Details about the location of the output of the batch inference job.
|
1335
|
+
*/
|
1336
|
+
outputDataConfig: ModelInvocationJobOutputDataConfig;
|
1337
|
+
/**
|
1338
|
+
* The number of hours after which batch inference job was set to time out.
|
1339
|
+
*/
|
1340
|
+
timeoutDurationInHours?: ModelInvocationJobTimeoutDurationInHours;
|
1341
|
+
/**
|
1342
|
+
* The time at which the batch inference job times or timed out.
|
1343
|
+
*/
|
1344
|
+
jobExpirationTime?: Timestamp;
|
1345
|
+
}
|
1210
1346
|
export interface GetModelInvocationLoggingConfigurationRequest {
|
1211
1347
|
}
|
1212
1348
|
export interface GetModelInvocationLoggingConfigurationResponse {
|
@@ -1886,6 +2022,50 @@ declare namespace Bedrock {
|
|
1886
2022
|
*/
|
1887
2023
|
modelCustomizationJobSummaries?: ModelCustomizationJobSummaries;
|
1888
2024
|
}
|
2025
|
+
export interface ListModelInvocationJobsRequest {
|
2026
|
+
/**
|
2027
|
+
* Specify a time to filter for batch inference jobs that were submitted after the time you specify.
|
2028
|
+
*/
|
2029
|
+
submitTimeAfter?: Timestamp;
|
2030
|
+
/**
|
2031
|
+
* Specify a time to filter for batch inference jobs that were submitted before the time you specify.
|
2032
|
+
*/
|
2033
|
+
submitTimeBefore?: Timestamp;
|
2034
|
+
/**
|
2035
|
+
* Specify a status to filter for batch inference jobs whose statuses match the string you specify.
|
2036
|
+
*/
|
2037
|
+
statusEquals?: ModelInvocationJobStatus;
|
2038
|
+
/**
|
2039
|
+
* Specify a string to filter for batch inference jobs whose names contain the string.
|
2040
|
+
*/
|
2041
|
+
nameContains?: ModelInvocationJobName;
|
2042
|
+
/**
|
2043
|
+
* The maximum number of results to return. If there are more results than the number that you specify, a nextToken value is returned. Use the nextToken in a request to return the next batch of results.
|
2044
|
+
*/
|
2045
|
+
maxResults?: MaxResults;
|
2046
|
+
/**
|
2047
|
+
* If there were more results than the value you specified in the maxResults field in a previous ListModelInvocationJobs request, the response would have returned a nextToken value. To see the next batch of results, send the nextToken value in another request.
|
2048
|
+
*/
|
2049
|
+
nextToken?: PaginationToken;
|
2050
|
+
/**
|
2051
|
+
* An attribute by which to sort the results.
|
2052
|
+
*/
|
2053
|
+
sortBy?: SortJobsBy;
|
2054
|
+
/**
|
2055
|
+
* Specifies whether to sort the results by ascending or descending order.
|
2056
|
+
*/
|
2057
|
+
sortOrder?: SortOrder;
|
2058
|
+
}
|
2059
|
+
export interface ListModelInvocationJobsResponse {
|
2060
|
+
/**
|
2061
|
+
* If there are more results than can fit in the response, a nextToken is returned. Use the nextToken in a request to return the next batch of results.
|
2062
|
+
*/
|
2063
|
+
nextToken?: PaginationToken;
|
2064
|
+
/**
|
2065
|
+
* A list of items, each of which contains a summary about a batch inference job.
|
2066
|
+
*/
|
2067
|
+
invocationJobSummaries?: ModelInvocationJobSummaries;
|
2068
|
+
}
|
1889
2069
|
export interface ListProvisionedModelThroughputsRequest {
|
1890
2070
|
/**
|
1891
2071
|
* A filter that returns Provisioned Throughputs created after the specified time.
|
@@ -1970,6 +2150,7 @@ declare namespace Bedrock {
|
|
1970
2150
|
embeddingDataDeliveryEnabled?: Boolean;
|
1971
2151
|
}
|
1972
2152
|
export type MaxResults = number;
|
2153
|
+
export type Message = string;
|
1973
2154
|
export type MetricFloat = number;
|
1974
2155
|
export type ModelArn = string;
|
1975
2156
|
export type ModelCopyJobArn = string;
|
@@ -2070,7 +2251,105 @@ declare namespace Bedrock {
|
|
2070
2251
|
customizationType?: CustomizationType;
|
2071
2252
|
}
|
2072
2253
|
export type ModelCustomizationList = ModelCustomization[];
|
2254
|
+
export type ModelId = string;
|
2073
2255
|
export type ModelIdentifier = string;
|
2256
|
+
export type ModelInvocationIdempotencyToken = string;
|
2257
|
+
export type ModelInvocationJobArn = string;
|
2258
|
+
export type ModelInvocationJobIdentifier = string;
|
2259
|
+
export interface ModelInvocationJobInputDataConfig {
|
2260
|
+
/**
|
2261
|
+
* Contains the configuration of the S3 location of the input data.
|
2262
|
+
*/
|
2263
|
+
s3InputDataConfig?: ModelInvocationJobS3InputDataConfig;
|
2264
|
+
}
|
2265
|
+
export type ModelInvocationJobName = string;
|
2266
|
+
export interface ModelInvocationJobOutputDataConfig {
|
2267
|
+
/**
|
2268
|
+
* Contains the configuration of the S3 location of the output data.
|
2269
|
+
*/
|
2270
|
+
s3OutputDataConfig?: ModelInvocationJobS3OutputDataConfig;
|
2271
|
+
}
|
2272
|
+
export interface ModelInvocationJobS3InputDataConfig {
|
2273
|
+
/**
|
2274
|
+
* The format of the input data.
|
2275
|
+
*/
|
2276
|
+
s3InputFormat?: S3InputFormat;
|
2277
|
+
/**
|
2278
|
+
* The S3 location of the input data.
|
2279
|
+
*/
|
2280
|
+
s3Uri: S3Uri;
|
2281
|
+
}
|
2282
|
+
export interface ModelInvocationJobS3OutputDataConfig {
|
2283
|
+
/**
|
2284
|
+
* The S3 location of the output data.
|
2285
|
+
*/
|
2286
|
+
s3Uri: S3Uri;
|
2287
|
+
/**
|
2288
|
+
* The unique identifier of the key that encrypts the S3 location of the output data.
|
2289
|
+
*/
|
2290
|
+
s3EncryptionKeyId?: KmsKeyId;
|
2291
|
+
}
|
2292
|
+
export type ModelInvocationJobStatus = "Submitted"|"InProgress"|"Completed"|"Failed"|"Stopping"|"Stopped"|"PartiallyCompleted"|"Expired"|"Validating"|"Scheduled"|string;
|
2293
|
+
export type ModelInvocationJobSummaries = ModelInvocationJobSummary[];
|
2294
|
+
export interface ModelInvocationJobSummary {
|
2295
|
+
/**
|
2296
|
+
* The Amazon Resource Name (ARN) of the batch inference job.
|
2297
|
+
*/
|
2298
|
+
jobArn: ModelInvocationJobArn;
|
2299
|
+
/**
|
2300
|
+
* The name of the batch inference job.
|
2301
|
+
*/
|
2302
|
+
jobName: ModelInvocationJobName;
|
2303
|
+
/**
|
2304
|
+
* The unique identifier of the foundation model used for model inference.
|
2305
|
+
*/
|
2306
|
+
modelId: ModelId;
|
2307
|
+
/**
|
2308
|
+
* A unique, case-sensitive identifier to ensure that the API request completes no more than one time. If this token matches a previous request, Amazon Bedrock ignores the request, but does not return an error. For more information, see Ensuring idempotency.
|
2309
|
+
*/
|
2310
|
+
clientRequestToken?: ModelInvocationIdempotencyToken;
|
2311
|
+
/**
|
2312
|
+
* The Amazon Resource Name (ARN) of the service role with permissions to carry out and manage batch inference. You can use the console to create a default service role or follow the steps at Create a service role for batch inference.
|
2313
|
+
*/
|
2314
|
+
roleArn: RoleArn;
|
2315
|
+
/**
|
2316
|
+
* The status of the batch inference job.
|
2317
|
+
*/
|
2318
|
+
status?: ModelInvocationJobStatus;
|
2319
|
+
/**
|
2320
|
+
* If the batch inference job failed, this field contains a message describing why the job failed.
|
2321
|
+
*/
|
2322
|
+
message?: Message;
|
2323
|
+
/**
|
2324
|
+
* The time at which the batch inference job was submitted.
|
2325
|
+
*/
|
2326
|
+
submitTime: Timestamp;
|
2327
|
+
/**
|
2328
|
+
* The time at which the batch inference job was last modified.
|
2329
|
+
*/
|
2330
|
+
lastModifiedTime?: Timestamp;
|
2331
|
+
/**
|
2332
|
+
* The time at which the batch inference job ended.
|
2333
|
+
*/
|
2334
|
+
endTime?: Timestamp;
|
2335
|
+
/**
|
2336
|
+
* Details about the location of the input to the batch inference job.
|
2337
|
+
*/
|
2338
|
+
inputDataConfig: ModelInvocationJobInputDataConfig;
|
2339
|
+
/**
|
2340
|
+
* Details about the location of the output of the batch inference job.
|
2341
|
+
*/
|
2342
|
+
outputDataConfig: ModelInvocationJobOutputDataConfig;
|
2343
|
+
/**
|
2344
|
+
* The number of hours after which the batch inference job was set to time out.
|
2345
|
+
*/
|
2346
|
+
timeoutDurationInHours?: ModelInvocationJobTimeoutDurationInHours;
|
2347
|
+
/**
|
2348
|
+
* The time at which the batch inference job times or timed out.
|
2349
|
+
*/
|
2350
|
+
jobExpirationTime?: Timestamp;
|
2351
|
+
}
|
2352
|
+
export type ModelInvocationJobTimeoutDurationInHours = number;
|
2074
2353
|
export type ModelModality = "TEXT"|"IMAGE"|"EMBEDDING"|string;
|
2075
2354
|
export type ModelModalityList = ModelModality[];
|
2076
2355
|
export type ModelName = string;
|
@@ -2157,6 +2436,7 @@ declare namespace Bedrock {
|
|
2157
2436
|
*/
|
2158
2437
|
keyPrefix?: KeyPrefix;
|
2159
2438
|
}
|
2439
|
+
export type S3InputFormat = "JSONL"|string;
|
2160
2440
|
export type S3Uri = string;
|
2161
2441
|
export type SageMakerFlowDefinitionArn = string;
|
2162
2442
|
export type SecurityGroupId = string;
|
@@ -2181,6 +2461,14 @@ declare namespace Bedrock {
|
|
2181
2461
|
}
|
2182
2462
|
export interface StopModelCustomizationJobResponse {
|
2183
2463
|
}
|
2464
|
+
export interface StopModelInvocationJobRequest {
|
2465
|
+
/**
|
2466
|
+
* The Amazon Resource Name (ARN) of the batch inference job to stop.
|
2467
|
+
*/
|
2468
|
+
jobIdentifier: ModelInvocationJobIdentifier;
|
2469
|
+
}
|
2470
|
+
export interface StopModelInvocationJobResponse {
|
2471
|
+
}
|
2184
2472
|
export type String = string;
|
2185
2473
|
export type SubnetId = string;
|
2186
2474
|
export type SubnetIds = SubnetId[];
|
@@ -1052,6 +1052,10 @@ declare namespace CodeBuild {
|
|
1052
1052
|
*/
|
1053
1053
|
overflowBehavior?: FleetOverflowBehavior;
|
1054
1054
|
vpcConfig?: VpcConfig;
|
1055
|
+
/**
|
1056
|
+
* The Amazon Machine Image (AMI) of the compute fleet.
|
1057
|
+
*/
|
1058
|
+
imageId?: NonEmptyString;
|
1055
1059
|
/**
|
1056
1060
|
* The service role associated with the compute fleet. For more information, see Allow a user to add a permission policy for a fleet service role in the CodeBuild User Guide.
|
1057
1061
|
*/
|
@@ -1415,7 +1419,7 @@ declare namespace CodeBuild {
|
|
1415
1419
|
languages?: EnvironmentLanguages;
|
1416
1420
|
}
|
1417
1421
|
export type EnvironmentPlatforms = EnvironmentPlatform[];
|
1418
|
-
export type EnvironmentType = "WINDOWS_CONTAINER"|"LINUX_CONTAINER"|"LINUX_GPU_CONTAINER"|"ARM_CONTAINER"|"WINDOWS_SERVER_2019_CONTAINER"|"LINUX_LAMBDA_CONTAINER"|"ARM_LAMBDA_CONTAINER"|string;
|
1422
|
+
export type EnvironmentType = "WINDOWS_CONTAINER"|"LINUX_CONTAINER"|"LINUX_GPU_CONTAINER"|"ARM_CONTAINER"|"WINDOWS_SERVER_2019_CONTAINER"|"LINUX_LAMBDA_CONTAINER"|"ARM_LAMBDA_CONTAINER"|"MAC_ARM"|string;
|
1419
1423
|
export interface EnvironmentVariable {
|
1420
1424
|
/**
|
1421
1425
|
* The name or key of the environment variable.
|
@@ -1492,6 +1496,10 @@ declare namespace CodeBuild {
|
|
1492
1496
|
*/
|
1493
1497
|
overflowBehavior?: FleetOverflowBehavior;
|
1494
1498
|
vpcConfig?: VpcConfig;
|
1499
|
+
/**
|
1500
|
+
* The Amazon Machine Image (AMI) of the compute fleet.
|
1501
|
+
*/
|
1502
|
+
imageId?: NonEmptyString;
|
1495
1503
|
/**
|
1496
1504
|
* The service role associated with the compute fleet. For more information, see Allow a user to add a permission policy for a fleet service role in the CodeBuild User Guide.
|
1497
1505
|
*/
|
@@ -1503,7 +1511,7 @@ declare namespace CodeBuild {
|
|
1503
1511
|
}
|
1504
1512
|
export type FleetArns = NonEmptyString[];
|
1505
1513
|
export type FleetCapacity = number;
|
1506
|
-
export type FleetContextCode = "CREATE_FAILED"|"UPDATE_FAILED"|"ACTION_REQUIRED"|string;
|
1514
|
+
export type FleetContextCode = "CREATE_FAILED"|"UPDATE_FAILED"|"ACTION_REQUIRED"|"PENDING_DELETION"|"INSUFFICIENT_CAPACITY"|string;
|
1507
1515
|
export type FleetName = string;
|
1508
1516
|
export type FleetNames = NonEmptyString[];
|
1509
1517
|
export type FleetOverflowBehavior = "QUEUE"|"ON_DEMAND"|string;
|
@@ -3059,6 +3067,10 @@ declare namespace CodeBuild {
|
|
3059
3067
|
*/
|
3060
3068
|
overflowBehavior?: FleetOverflowBehavior;
|
3061
3069
|
vpcConfig?: VpcConfig;
|
3070
|
+
/**
|
3071
|
+
* The Amazon Machine Image (AMI) of the compute fleet.
|
3072
|
+
*/
|
3073
|
+
imageId?: NonEmptyString;
|
3062
3074
|
/**
|
3063
3075
|
* The service role associated with the compute fleet. For more information, see Allow a user to add a permission policy for a fleet service role in the CodeBuild User Guide.
|
3064
3076
|
*/
|