cdk-lambda-subminute 2.0.482 → 2.0.484
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.jsii +4 -4
- package/API.md +494 -80
- package/lib/cdk-lambda-subminute.js +3 -3
- package/node_modules/aws-sdk/README.md +1 -1
- package/node_modules/aws-sdk/apis/bedrock-2023-04-20.min.json +123 -9
- package/node_modules/aws-sdk/apis/bedrock-2023-04-20.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/chatbot-2017-10-11.min.json +73 -24
- package/node_modules/aws-sdk/apis/chatbot-2017-10-11.paginators.json +14 -7
- package/node_modules/aws-sdk/apis/iotsitewise-2019-12-02.min.json +247 -161
- package/node_modules/aws-sdk/apis/omics-2022-11-28.min.json +47 -36
- package/node_modules/aws-sdk/apis/workspaces-2015-04-08.min.json +78 -40
- package/node_modules/aws-sdk/clients/bedrock.d.ts +135 -2
- package/node_modules/aws-sdk/clients/chatbot.d.ts +154 -155
- package/node_modules/aws-sdk/clients/iotsitewise.d.ts +169 -72
- package/node_modules/aws-sdk/clients/omics.d.ts +32 -10
- package/node_modules/aws-sdk/clients/polly.d.ts +2 -2
- package/node_modules/aws-sdk/clients/workspaces.d.ts +57 -2
- package/node_modules/aws-sdk/dist/aws-sdk-core-react-native.js +1 -1
- package/node_modules/aws-sdk/dist/aws-sdk-react-native.js +8 -8
- package/node_modules/aws-sdk/dist/aws-sdk.js +3 -3
- package/node_modules/aws-sdk/dist/aws-sdk.min.js +2 -2
- package/node_modules/aws-sdk/lib/core.js +1 -1
- package/node_modules/aws-sdk/package.json +1 -1
- package/package.json +7 -7
@@ -68,11 +68,11 @@ declare class Bedrock extends Service {
|
|
68
68
|
*/
|
69
69
|
createModelImportJob(callback?: (err: AWSError, data: Bedrock.Types.CreateModelImportJobResponse) => void): Request<Bedrock.Types.CreateModelImportJobResponse, AWSError>;
|
70
70
|
/**
|
71
|
-
* Creates a job to invoke a model on multiple prompts
|
71
|
+
* Creates a batch inference job to invoke a model on multiple prompts. Format your data according to Format your inference data and upload it to an Amazon S3 bucket. For more information, see Process multiple prompts with batch inference. The response returns a jobArn that you can use to stop or get details about the job.
|
72
72
|
*/
|
73
73
|
createModelInvocationJob(params: Bedrock.Types.CreateModelInvocationJobRequest, callback?: (err: AWSError, data: Bedrock.Types.CreateModelInvocationJobResponse) => void): Request<Bedrock.Types.CreateModelInvocationJobResponse, AWSError>;
|
74
74
|
/**
|
75
|
-
* Creates a job to invoke a model on multiple prompts
|
75
|
+
* Creates a batch inference job to invoke a model on multiple prompts. Format your data according to Format your inference data and upload it to an Amazon S3 bucket. For more information, see Process multiple prompts with batch inference. The response returns a jobArn that you can use to stop or get details about the job.
|
76
76
|
*/
|
77
77
|
createModelInvocationJob(callback?: (err: AWSError, data: Bedrock.Types.CreateModelInvocationJobResponse) => void): Request<Bedrock.Types.CreateModelInvocationJobResponse, AWSError>;
|
78
78
|
/**
|
@@ -163,6 +163,14 @@ declare class Bedrock extends Service {
|
|
163
163
|
* Gets properties associated with a customized model you imported.
|
164
164
|
*/
|
165
165
|
getImportedModel(callback?: (err: AWSError, data: Bedrock.Types.GetImportedModelResponse) => void): Request<Bedrock.Types.GetImportedModelResponse, AWSError>;
|
166
|
+
/**
|
167
|
+
* Gets information about an inference profile. For more information, see the Amazon Bedrock User Guide.
|
168
|
+
*/
|
169
|
+
getInferenceProfile(params: Bedrock.Types.GetInferenceProfileRequest, callback?: (err: AWSError, data: Bedrock.Types.GetInferenceProfileResponse) => void): Request<Bedrock.Types.GetInferenceProfileResponse, AWSError>;
|
170
|
+
/**
|
171
|
+
* Gets information about an inference profile. For more information, see the Amazon Bedrock User Guide.
|
172
|
+
*/
|
173
|
+
getInferenceProfile(callback?: (err: AWSError, data: Bedrock.Types.GetInferenceProfileResponse) => void): Request<Bedrock.Types.GetInferenceProfileResponse, AWSError>;
|
166
174
|
/**
|
167
175
|
* Retrieves information about a model copy job. For more information, see Copy models to be used in other regions in the Amazon Bedrock User Guide.
|
168
176
|
*/
|
@@ -251,6 +259,14 @@ declare class Bedrock extends Service {
|
|
251
259
|
* Returns a list of models you've imported. You can filter the results to return based on one or more criteria. For more information, see Import a customized model in the Amazon Bedrock User Guide.
|
252
260
|
*/
|
253
261
|
listImportedModels(callback?: (err: AWSError, data: Bedrock.Types.ListImportedModelsResponse) => void): Request<Bedrock.Types.ListImportedModelsResponse, AWSError>;
|
262
|
+
/**
|
263
|
+
* Returns a list of inference profiles that you can use.
|
264
|
+
*/
|
265
|
+
listInferenceProfiles(params: Bedrock.Types.ListInferenceProfilesRequest, callback?: (err: AWSError, data: Bedrock.Types.ListInferenceProfilesResponse) => void): Request<Bedrock.Types.ListInferenceProfilesResponse, AWSError>;
|
266
|
+
/**
|
267
|
+
* Returns a list of inference profiles that you can use.
|
268
|
+
*/
|
269
|
+
listInferenceProfiles(callback?: (err: AWSError, data: Bedrock.Types.ListInferenceProfilesResponse) => void): Request<Bedrock.Types.ListInferenceProfilesResponse, AWSError>;
|
254
270
|
/**
|
255
271
|
* Returns a list of model copy jobs that you have submitted. You can filter the jobs to return based on one or more criteria. For more information, see Copy models to be used in other regions in the Amazon Bedrock User Guide.
|
256
272
|
*/
|
@@ -1326,6 +1342,50 @@ declare namespace Bedrock {
|
|
1326
1342
|
*/
|
1327
1343
|
modelKmsKeyArn?: KmsKeyArn;
|
1328
1344
|
}
|
1345
|
+
export interface GetInferenceProfileRequest {
|
1346
|
+
/**
|
1347
|
+
* The unique identifier of the inference profile.
|
1348
|
+
*/
|
1349
|
+
inferenceProfileIdentifier: InferenceProfileIdentifier;
|
1350
|
+
}
|
1351
|
+
export interface GetInferenceProfileResponse {
|
1352
|
+
/**
|
1353
|
+
* The name of the inference profile.
|
1354
|
+
*/
|
1355
|
+
inferenceProfileName: InferenceProfileName;
|
1356
|
+
/**
|
1357
|
+
* A list of information about each model in the inference profile.
|
1358
|
+
*/
|
1359
|
+
models: InferenceProfileModels;
|
1360
|
+
/**
|
1361
|
+
* The description of the inference profile.
|
1362
|
+
*/
|
1363
|
+
description?: InferenceProfileDescription;
|
1364
|
+
/**
|
1365
|
+
* The time at which the inference profile was created.
|
1366
|
+
*/
|
1367
|
+
createdAt?: Timestamp;
|
1368
|
+
/**
|
1369
|
+
* The time at which the inference profile was last updated.
|
1370
|
+
*/
|
1371
|
+
updatedAt?: Timestamp;
|
1372
|
+
/**
|
1373
|
+
* The Amazon Resource Name (ARN) of the inference profile.
|
1374
|
+
*/
|
1375
|
+
inferenceProfileArn: InferenceProfileArn;
|
1376
|
+
/**
|
1377
|
+
* The unique identifier of the inference profile.
|
1378
|
+
*/
|
1379
|
+
inferenceProfileId: InferenceProfileId;
|
1380
|
+
/**
|
1381
|
+
* The status of the inference profile. ACTIVE means that the inference profile is available to use.
|
1382
|
+
*/
|
1383
|
+
status: InferenceProfileStatus;
|
1384
|
+
/**
|
1385
|
+
* The type of the inference profile. SYSTEM_DEFINED means that the inference profile is defined by Amazon Bedrock.
|
1386
|
+
*/
|
1387
|
+
type: InferenceProfileType;
|
1388
|
+
}
|
1329
1389
|
export interface GetModelCopyJobRequest {
|
1330
1390
|
/**
|
1331
1391
|
* The Amazon Resource Name (ARN) of the model copy job.
|
@@ -2045,6 +2105,59 @@ declare namespace Bedrock {
|
|
2045
2105
|
creationTime: Timestamp;
|
2046
2106
|
}
|
2047
2107
|
export type ImportedModelSummaryList = ImportedModelSummary[];
|
2108
|
+
export type InferenceProfileArn = string;
|
2109
|
+
export type InferenceProfileDescription = string;
|
2110
|
+
export type InferenceProfileId = string;
|
2111
|
+
export type InferenceProfileIdentifier = string;
|
2112
|
+
export interface InferenceProfileModel {
|
2113
|
+
/**
|
2114
|
+
* The Amazon Resource Name (ARN) of the model.
|
2115
|
+
*/
|
2116
|
+
modelArn?: FoundationModelArn;
|
2117
|
+
}
|
2118
|
+
export type InferenceProfileModels = InferenceProfileModel[];
|
2119
|
+
export type InferenceProfileName = string;
|
2120
|
+
export type InferenceProfileStatus = "ACTIVE"|string;
|
2121
|
+
export type InferenceProfileSummaries = InferenceProfileSummary[];
|
2122
|
+
export interface InferenceProfileSummary {
|
2123
|
+
/**
|
2124
|
+
* The name of the inference profile.
|
2125
|
+
*/
|
2126
|
+
inferenceProfileName: InferenceProfileName;
|
2127
|
+
/**
|
2128
|
+
* A list of information about each model in the inference profile.
|
2129
|
+
*/
|
2130
|
+
models: InferenceProfileModels;
|
2131
|
+
/**
|
2132
|
+
* The description of the inference profile.
|
2133
|
+
*/
|
2134
|
+
description?: InferenceProfileDescription;
|
2135
|
+
/**
|
2136
|
+
* The time at which the inference profile was created.
|
2137
|
+
*/
|
2138
|
+
createdAt?: Timestamp;
|
2139
|
+
/**
|
2140
|
+
* The time at which the inference profile was last updated.
|
2141
|
+
*/
|
2142
|
+
updatedAt?: Timestamp;
|
2143
|
+
/**
|
2144
|
+
* The Amazon Resource Name (ARN) of the inference profile.
|
2145
|
+
*/
|
2146
|
+
inferenceProfileArn: InferenceProfileArn;
|
2147
|
+
/**
|
2148
|
+
* The unique identifier of the inference profile.
|
2149
|
+
*/
|
2150
|
+
inferenceProfileId: InferenceProfileId;
|
2151
|
+
/**
|
2152
|
+
* The status of the inference profile. ACTIVE means that the inference profile is available to use.
|
2153
|
+
*/
|
2154
|
+
status: InferenceProfileStatus;
|
2155
|
+
/**
|
2156
|
+
* The type of the inference profile. SYSTEM_DEFINED means that the inference profile is defined by Amazon Bedrock.
|
2157
|
+
*/
|
2158
|
+
type: InferenceProfileType;
|
2159
|
+
}
|
2160
|
+
export type InferenceProfileType = "SYSTEM_DEFINED"|string;
|
2048
2161
|
export type InferenceType = "ON_DEMAND"|"PROVISIONED"|string;
|
2049
2162
|
export type InferenceTypeList = InferenceType[];
|
2050
2163
|
export type JobName = string;
|
@@ -2235,6 +2348,26 @@ declare namespace Bedrock {
|
|
2235
2348
|
*/
|
2236
2349
|
modelSummaries?: ImportedModelSummaryList;
|
2237
2350
|
}
|
2351
|
+
export interface ListInferenceProfilesRequest {
|
2352
|
+
/**
|
2353
|
+
* The maximum number of results to return in the response. If the total number of results is greater than this value, use the token returned in the response in the nextToken field when making another request to return the next batch of results.
|
2354
|
+
*/
|
2355
|
+
maxResults?: MaxResults;
|
2356
|
+
/**
|
2357
|
+
* If the total number of results is greater than the maxResults value provided in the request, enter the token returned in the nextToken field in the response in this field to return the next batch of results.
|
2358
|
+
*/
|
2359
|
+
nextToken?: PaginationToken;
|
2360
|
+
}
|
2361
|
+
export interface ListInferenceProfilesResponse {
|
2362
|
+
/**
|
2363
|
+
* A list of information about each inference profile that you can use.
|
2364
|
+
*/
|
2365
|
+
inferenceProfileSummaries?: InferenceProfileSummaries;
|
2366
|
+
/**
|
2367
|
+
* If the total number of results is greater than the maxResults value provided in the request, use this token when making another request in the nextToken field to return the next batch of results.
|
2368
|
+
*/
|
2369
|
+
nextToken?: PaginationToken;
|
2370
|
+
}
|
2238
2371
|
export interface ListModelCopyJobsRequest {
|
2239
2372
|
/**
|
2240
2373
|
* Filters for model copy jobs created after the specified time.
|