aws-sdk 2.1683.0 → 2.1684.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +1 -1
- package/apis/bedrock-2023-04-20.min.json +123 -9
- package/apis/bedrock-2023-04-20.paginators.json +6 -0
- package/apis/chatbot-2017-10-11.min.json +73 -24
- package/apis/chatbot-2017-10-11.paginators.json +14 -7
- package/apis/omics-2022-11-28.min.json +47 -36
- package/clients/bedrock.d.ts +135 -2
- package/clients/chatbot.d.ts +154 -155
- package/clients/omics.d.ts +32 -10
- package/clients/polly.d.ts +2 -2
- package/dist/aws-sdk-core-react-native.js +1 -1
- package/dist/aws-sdk-react-native.js +6 -6
- package/dist/aws-sdk.js +3 -3
- package/dist/aws-sdk.min.js +2 -2
- package/lib/core.js +1 -1
- package/package.json +1 -1
@@ -2,9 +2,14 @@
|
|
2
2
|
"version": "2.0",
|
3
3
|
"metadata": {
|
4
4
|
"apiVersion": "2022-11-28",
|
5
|
+
"auth": [
|
6
|
+
"aws.auth#sigv4"
|
7
|
+
],
|
5
8
|
"endpointPrefix": "omics",
|
6
|
-
"jsonVersion": "1.1",
|
7
9
|
"protocol": "rest-json",
|
10
|
+
"protocols": [
|
11
|
+
"rest-json"
|
12
|
+
],
|
8
13
|
"serviceFullName": "Amazon Omics",
|
9
14
|
"serviceId": "Omics",
|
10
15
|
"signatureVersion": "v4",
|
@@ -1469,7 +1474,8 @@
|
|
1469
1474
|
"description": {},
|
1470
1475
|
"tags": {
|
1471
1476
|
"shape": "Sy"
|
1472
|
-
}
|
1477
|
+
},
|
1478
|
+
"readSetId": {}
|
1473
1479
|
}
|
1474
1480
|
}
|
1475
1481
|
}
|
@@ -1547,7 +1553,8 @@
|
|
1547
1553
|
"creationType": {},
|
1548
1554
|
"etag": {
|
1549
1555
|
"shape": "S60"
|
1550
|
-
}
|
1556
|
+
},
|
1557
|
+
"creationJobId": {}
|
1551
1558
|
}
|
1552
1559
|
},
|
1553
1560
|
"endpoint": {
|
@@ -1665,7 +1672,8 @@
|
|
1665
1672
|
"description": {},
|
1666
1673
|
"tags": {
|
1667
1674
|
"shape": "Sy"
|
1668
|
-
}
|
1675
|
+
},
|
1676
|
+
"referenceId": {}
|
1669
1677
|
}
|
1670
1678
|
}
|
1671
1679
|
}
|
@@ -1732,7 +1740,9 @@
|
|
1732
1740
|
"shape": "S5t"
|
1733
1741
|
}
|
1734
1742
|
}
|
1735
|
-
}
|
1743
|
+
},
|
1744
|
+
"creationType": {},
|
1745
|
+
"creationJobId": {}
|
1736
1746
|
}
|
1737
1747
|
},
|
1738
1748
|
"endpoint": {
|
@@ -1823,7 +1833,7 @@
|
|
1823
1833
|
"definition": {},
|
1824
1834
|
"digest": {},
|
1825
1835
|
"parameters": {
|
1826
|
-
"shape": "
|
1836
|
+
"shape": "S70"
|
1827
1837
|
},
|
1828
1838
|
"storageCapacity": {
|
1829
1839
|
"type": "integer"
|
@@ -1837,13 +1847,13 @@
|
|
1837
1847
|
},
|
1838
1848
|
"startedBy": {},
|
1839
1849
|
"creationTime": {
|
1840
|
-
"shape": "
|
1850
|
+
"shape": "S78"
|
1841
1851
|
},
|
1842
1852
|
"startTime": {
|
1843
|
-
"shape": "
|
1853
|
+
"shape": "S78"
|
1844
1854
|
},
|
1845
1855
|
"stopTime": {
|
1846
|
-
"shape": "
|
1856
|
+
"shape": "S78"
|
1847
1857
|
},
|
1848
1858
|
"statusMessage": {},
|
1849
1859
|
"tags": {
|
@@ -1903,7 +1913,7 @@
|
|
1903
1913
|
"type": "integer"
|
1904
1914
|
},
|
1905
1915
|
"creationTime": {
|
1906
|
-
"shape": "
|
1916
|
+
"shape": "S7n"
|
1907
1917
|
},
|
1908
1918
|
"tags": {
|
1909
1919
|
"shape": "Sy"
|
@@ -1953,13 +1963,13 @@
|
|
1953
1963
|
"type": "integer"
|
1954
1964
|
},
|
1955
1965
|
"creationTime": {
|
1956
|
-
"shape": "
|
1966
|
+
"shape": "S7w"
|
1957
1967
|
},
|
1958
1968
|
"startTime": {
|
1959
|
-
"shape": "
|
1969
|
+
"shape": "S7w"
|
1960
1970
|
},
|
1961
1971
|
"stopTime": {
|
1962
|
-
"shape": "
|
1972
|
+
"shape": "S7w"
|
1963
1973
|
},
|
1964
1974
|
"statusMessage": {},
|
1965
1975
|
"logStream": {},
|
@@ -2047,7 +2057,7 @@
|
|
2047
2057
|
"type": "structure",
|
2048
2058
|
"members": {
|
2049
2059
|
"share": {
|
2050
|
-
"shape": "
|
2060
|
+
"shape": "S88"
|
2051
2061
|
}
|
2052
2062
|
}
|
2053
2063
|
},
|
@@ -2245,14 +2255,14 @@
|
|
2245
2255
|
"type": "integer"
|
2246
2256
|
},
|
2247
2257
|
"creationTime": {
|
2248
|
-
"shape": "
|
2258
|
+
"shape": "S8k"
|
2249
2259
|
},
|
2250
2260
|
"statusMessage": {},
|
2251
2261
|
"tags": {
|
2252
2262
|
"shape": "Sy"
|
2253
2263
|
},
|
2254
2264
|
"metadata": {
|
2255
|
-
"shape": "
|
2265
|
+
"shape": "S8m"
|
2256
2266
|
},
|
2257
2267
|
"accelerators": {}
|
2258
2268
|
}
|
@@ -3229,7 +3239,7 @@
|
|
3229
3239
|
"type": "integer"
|
3230
3240
|
},
|
3231
3241
|
"creationTime": {
|
3232
|
-
"shape": "
|
3242
|
+
"shape": "S7n"
|
3233
3243
|
},
|
3234
3244
|
"maxGpus": {
|
3235
3245
|
"type": "integer"
|
@@ -3293,13 +3303,13 @@
|
|
3293
3303
|
"type": "integer"
|
3294
3304
|
},
|
3295
3305
|
"creationTime": {
|
3296
|
-
"shape": "
|
3306
|
+
"shape": "S7w"
|
3297
3307
|
},
|
3298
3308
|
"startTime": {
|
3299
|
-
"shape": "
|
3309
|
+
"shape": "S7w"
|
3300
3310
|
},
|
3301
3311
|
"stopTime": {
|
3302
|
-
"shape": "
|
3312
|
+
"shape": "S7w"
|
3303
3313
|
},
|
3304
3314
|
"gpus": {
|
3305
3315
|
"type": "integer"
|
@@ -3367,13 +3377,13 @@
|
|
3367
3377
|
"type": "integer"
|
3368
3378
|
},
|
3369
3379
|
"creationTime": {
|
3370
|
-
"shape": "
|
3380
|
+
"shape": "S78"
|
3371
3381
|
},
|
3372
3382
|
"startTime": {
|
3373
|
-
"shape": "
|
3383
|
+
"shape": "S78"
|
3374
3384
|
},
|
3375
3385
|
"stopTime": {
|
3376
|
-
"shape": "
|
3386
|
+
"shape": "S78"
|
3377
3387
|
},
|
3378
3388
|
"storageType": {}
|
3379
3389
|
}
|
@@ -3504,7 +3514,7 @@
|
|
3504
3514
|
"shares": {
|
3505
3515
|
"type": "list",
|
3506
3516
|
"member": {
|
3507
|
-
"shape": "
|
3517
|
+
"shape": "S88"
|
3508
3518
|
}
|
3509
3519
|
},
|
3510
3520
|
"nextToken": {}
|
@@ -3746,10 +3756,10 @@
|
|
3746
3756
|
"type": {},
|
3747
3757
|
"digest": {},
|
3748
3758
|
"creationTime": {
|
3749
|
-
"shape": "
|
3759
|
+
"shape": "S8k"
|
3750
3760
|
},
|
3751
3761
|
"metadata": {
|
3752
|
-
"shape": "
|
3762
|
+
"shape": "S8m"
|
3753
3763
|
}
|
3754
3764
|
}
|
3755
3765
|
}
|
@@ -4078,7 +4088,7 @@
|
|
4078
4088
|
"type": "integer"
|
4079
4089
|
},
|
4080
4090
|
"parameters": {
|
4081
|
-
"shape": "
|
4091
|
+
"shape": "S70"
|
4082
4092
|
},
|
4083
4093
|
"storageCapacity": {
|
4084
4094
|
"type": "integer"
|
@@ -4489,10 +4499,11 @@
|
|
4489
4499
|
"checksum": {}
|
4490
4500
|
}
|
4491
4501
|
},
|
4492
|
-
"authtype": "v4-unsigned-body",
|
4493
4502
|
"endpoint": {
|
4494
4503
|
"hostPrefix": "storage-"
|
4495
|
-
}
|
4504
|
+
},
|
4505
|
+
"unsignedPayload": true,
|
4506
|
+
"authtype": "v4-unsigned-body"
|
4496
4507
|
}
|
4497
4508
|
},
|
4498
4509
|
"shapes": {
|
@@ -4697,24 +4708,24 @@
|
|
4697
4708
|
"source2": {}
|
4698
4709
|
}
|
4699
4710
|
},
|
4700
|
-
"
|
4711
|
+
"S70": {
|
4701
4712
|
"type": "structure",
|
4702
4713
|
"members": {},
|
4703
4714
|
"document": true
|
4704
4715
|
},
|
4705
|
-
"
|
4716
|
+
"S78": {
|
4706
4717
|
"type": "timestamp",
|
4707
4718
|
"timestampFormat": "iso8601"
|
4708
4719
|
},
|
4709
|
-
"
|
4720
|
+
"S7n": {
|
4710
4721
|
"type": "timestamp",
|
4711
4722
|
"timestampFormat": "iso8601"
|
4712
4723
|
},
|
4713
|
-
"
|
4724
|
+
"S7w": {
|
4714
4725
|
"type": "timestamp",
|
4715
4726
|
"timestampFormat": "iso8601"
|
4716
4727
|
},
|
4717
|
-
"
|
4728
|
+
"S88": {
|
4718
4729
|
"type": "structure",
|
4719
4730
|
"members": {
|
4720
4731
|
"shareId": {},
|
@@ -4733,11 +4744,11 @@
|
|
4733
4744
|
}
|
4734
4745
|
}
|
4735
4746
|
},
|
4736
|
-
"
|
4747
|
+
"S8k": {
|
4737
4748
|
"type": "timestamp",
|
4738
4749
|
"timestampFormat": "iso8601"
|
4739
4750
|
},
|
4740
|
-
"
|
4751
|
+
"S8m": {
|
4741
4752
|
"type": "map",
|
4742
4753
|
"key": {},
|
4743
4754
|
"value": {}
|
package/clients/bedrock.d.ts
CHANGED
@@ -68,11 +68,11 @@ declare class Bedrock extends Service {
|
|
68
68
|
*/
|
69
69
|
createModelImportJob(callback?: (err: AWSError, data: Bedrock.Types.CreateModelImportJobResponse) => void): Request<Bedrock.Types.CreateModelImportJobResponse, AWSError>;
|
70
70
|
/**
|
71
|
-
* Creates a job to invoke a model on multiple prompts
|
71
|
+
* Creates a batch inference job to invoke a model on multiple prompts. Format your data according to Format your inference data and upload it to an Amazon S3 bucket. For more information, see Process multiple prompts with batch inference. The response returns a jobArn that you can use to stop or get details about the job.
|
72
72
|
*/
|
73
73
|
createModelInvocationJob(params: Bedrock.Types.CreateModelInvocationJobRequest, callback?: (err: AWSError, data: Bedrock.Types.CreateModelInvocationJobResponse) => void): Request<Bedrock.Types.CreateModelInvocationJobResponse, AWSError>;
|
74
74
|
/**
|
75
|
-
* Creates a job to invoke a model on multiple prompts
|
75
|
+
* Creates a batch inference job to invoke a model on multiple prompts. Format your data according to Format your inference data and upload it to an Amazon S3 bucket. For more information, see Process multiple prompts with batch inference. The response returns a jobArn that you can use to stop or get details about the job.
|
76
76
|
*/
|
77
77
|
createModelInvocationJob(callback?: (err: AWSError, data: Bedrock.Types.CreateModelInvocationJobResponse) => void): Request<Bedrock.Types.CreateModelInvocationJobResponse, AWSError>;
|
78
78
|
/**
|
@@ -163,6 +163,14 @@ declare class Bedrock extends Service {
|
|
163
163
|
* Gets properties associated with a customized model you imported.
|
164
164
|
*/
|
165
165
|
getImportedModel(callback?: (err: AWSError, data: Bedrock.Types.GetImportedModelResponse) => void): Request<Bedrock.Types.GetImportedModelResponse, AWSError>;
|
166
|
+
/**
|
167
|
+
* Gets information about an inference profile. For more information, see the Amazon Bedrock User Guide.
|
168
|
+
*/
|
169
|
+
getInferenceProfile(params: Bedrock.Types.GetInferenceProfileRequest, callback?: (err: AWSError, data: Bedrock.Types.GetInferenceProfileResponse) => void): Request<Bedrock.Types.GetInferenceProfileResponse, AWSError>;
|
170
|
+
/**
|
171
|
+
* Gets information about an inference profile. For more information, see the Amazon Bedrock User Guide.
|
172
|
+
*/
|
173
|
+
getInferenceProfile(callback?: (err: AWSError, data: Bedrock.Types.GetInferenceProfileResponse) => void): Request<Bedrock.Types.GetInferenceProfileResponse, AWSError>;
|
166
174
|
/**
|
167
175
|
* Retrieves information about a model copy job. For more information, see Copy models to be used in other regions in the Amazon Bedrock User Guide.
|
168
176
|
*/
|
@@ -251,6 +259,14 @@ declare class Bedrock extends Service {
|
|
251
259
|
* Returns a list of models you've imported. You can filter the results to return based on one or more criteria. For more information, see Import a customized model in the Amazon Bedrock User Guide.
|
252
260
|
*/
|
253
261
|
listImportedModels(callback?: (err: AWSError, data: Bedrock.Types.ListImportedModelsResponse) => void): Request<Bedrock.Types.ListImportedModelsResponse, AWSError>;
|
262
|
+
/**
|
263
|
+
* Returns a list of inference profiles that you can use.
|
264
|
+
*/
|
265
|
+
listInferenceProfiles(params: Bedrock.Types.ListInferenceProfilesRequest, callback?: (err: AWSError, data: Bedrock.Types.ListInferenceProfilesResponse) => void): Request<Bedrock.Types.ListInferenceProfilesResponse, AWSError>;
|
266
|
+
/**
|
267
|
+
* Returns a list of inference profiles that you can use.
|
268
|
+
*/
|
269
|
+
listInferenceProfiles(callback?: (err: AWSError, data: Bedrock.Types.ListInferenceProfilesResponse) => void): Request<Bedrock.Types.ListInferenceProfilesResponse, AWSError>;
|
254
270
|
/**
|
255
271
|
* Returns a list of model copy jobs that you have submitted. You can filter the jobs to return based on one or more criteria. For more information, see Copy models to be used in other regions in the Amazon Bedrock User Guide.
|
256
272
|
*/
|
@@ -1326,6 +1342,50 @@ declare namespace Bedrock {
|
|
1326
1342
|
*/
|
1327
1343
|
modelKmsKeyArn?: KmsKeyArn;
|
1328
1344
|
}
|
1345
|
+
export interface GetInferenceProfileRequest {
|
1346
|
+
/**
|
1347
|
+
* The unique identifier of the inference profile.
|
1348
|
+
*/
|
1349
|
+
inferenceProfileIdentifier: InferenceProfileIdentifier;
|
1350
|
+
}
|
1351
|
+
export interface GetInferenceProfileResponse {
|
1352
|
+
/**
|
1353
|
+
* The name of the inference profile.
|
1354
|
+
*/
|
1355
|
+
inferenceProfileName: InferenceProfileName;
|
1356
|
+
/**
|
1357
|
+
* A list of information about each model in the inference profile.
|
1358
|
+
*/
|
1359
|
+
models: InferenceProfileModels;
|
1360
|
+
/**
|
1361
|
+
* The description of the inference profile.
|
1362
|
+
*/
|
1363
|
+
description?: InferenceProfileDescription;
|
1364
|
+
/**
|
1365
|
+
* The time at which the inference profile was created.
|
1366
|
+
*/
|
1367
|
+
createdAt?: Timestamp;
|
1368
|
+
/**
|
1369
|
+
* The time at which the inference profile was last updated.
|
1370
|
+
*/
|
1371
|
+
updatedAt?: Timestamp;
|
1372
|
+
/**
|
1373
|
+
* The Amazon Resource Name (ARN) of the inference profile.
|
1374
|
+
*/
|
1375
|
+
inferenceProfileArn: InferenceProfileArn;
|
1376
|
+
/**
|
1377
|
+
* The unique identifier of the inference profile.
|
1378
|
+
*/
|
1379
|
+
inferenceProfileId: InferenceProfileId;
|
1380
|
+
/**
|
1381
|
+
* The status of the inference profile. ACTIVE means that the inference profile is available to use.
|
1382
|
+
*/
|
1383
|
+
status: InferenceProfileStatus;
|
1384
|
+
/**
|
1385
|
+
* The type of the inference profile. SYSTEM_DEFINED means that the inference profile is defined by Amazon Bedrock.
|
1386
|
+
*/
|
1387
|
+
type: InferenceProfileType;
|
1388
|
+
}
|
1329
1389
|
export interface GetModelCopyJobRequest {
|
1330
1390
|
/**
|
1331
1391
|
* The Amazon Resource Name (ARN) of the model copy job.
|
@@ -2045,6 +2105,59 @@ declare namespace Bedrock {
|
|
2045
2105
|
creationTime: Timestamp;
|
2046
2106
|
}
|
2047
2107
|
export type ImportedModelSummaryList = ImportedModelSummary[];
|
2108
|
+
export type InferenceProfileArn = string;
|
2109
|
+
export type InferenceProfileDescription = string;
|
2110
|
+
export type InferenceProfileId = string;
|
2111
|
+
export type InferenceProfileIdentifier = string;
|
2112
|
+
export interface InferenceProfileModel {
|
2113
|
+
/**
|
2114
|
+
* The Amazon Resource Name (ARN) of the model.
|
2115
|
+
*/
|
2116
|
+
modelArn?: FoundationModelArn;
|
2117
|
+
}
|
2118
|
+
export type InferenceProfileModels = InferenceProfileModel[];
|
2119
|
+
export type InferenceProfileName = string;
|
2120
|
+
export type InferenceProfileStatus = "ACTIVE"|string;
|
2121
|
+
export type InferenceProfileSummaries = InferenceProfileSummary[];
|
2122
|
+
export interface InferenceProfileSummary {
|
2123
|
+
/**
|
2124
|
+
* The name of the inference profile.
|
2125
|
+
*/
|
2126
|
+
inferenceProfileName: InferenceProfileName;
|
2127
|
+
/**
|
2128
|
+
* A list of information about each model in the inference profile.
|
2129
|
+
*/
|
2130
|
+
models: InferenceProfileModels;
|
2131
|
+
/**
|
2132
|
+
* The description of the inference profile.
|
2133
|
+
*/
|
2134
|
+
description?: InferenceProfileDescription;
|
2135
|
+
/**
|
2136
|
+
* The time at which the inference profile was created.
|
2137
|
+
*/
|
2138
|
+
createdAt?: Timestamp;
|
2139
|
+
/**
|
2140
|
+
* The time at which the inference profile was last updated.
|
2141
|
+
*/
|
2142
|
+
updatedAt?: Timestamp;
|
2143
|
+
/**
|
2144
|
+
* The Amazon Resource Name (ARN) of the inference profile.
|
2145
|
+
*/
|
2146
|
+
inferenceProfileArn: InferenceProfileArn;
|
2147
|
+
/**
|
2148
|
+
* The unique identifier of the inference profile.
|
2149
|
+
*/
|
2150
|
+
inferenceProfileId: InferenceProfileId;
|
2151
|
+
/**
|
2152
|
+
* The status of the inference profile. ACTIVE means that the inference profile is available to use.
|
2153
|
+
*/
|
2154
|
+
status: InferenceProfileStatus;
|
2155
|
+
/**
|
2156
|
+
* The type of the inference profile. SYSTEM_DEFINED means that the inference profile is defined by Amazon Bedrock.
|
2157
|
+
*/
|
2158
|
+
type: InferenceProfileType;
|
2159
|
+
}
|
2160
|
+
export type InferenceProfileType = "SYSTEM_DEFINED"|string;
|
2048
2161
|
export type InferenceType = "ON_DEMAND"|"PROVISIONED"|string;
|
2049
2162
|
export type InferenceTypeList = InferenceType[];
|
2050
2163
|
export type JobName = string;
|
@@ -2235,6 +2348,26 @@ declare namespace Bedrock {
|
|
2235
2348
|
*/
|
2236
2349
|
modelSummaries?: ImportedModelSummaryList;
|
2237
2350
|
}
|
2351
|
+
export interface ListInferenceProfilesRequest {
|
2352
|
+
/**
|
2353
|
+
* The maximum number of results to return in the response. If the total number of results is greater than this value, use the token returned in the response in the nextToken field when making another request to return the next batch of results.
|
2354
|
+
*/
|
2355
|
+
maxResults?: MaxResults;
|
2356
|
+
/**
|
2357
|
+
* If the total number of results is greater than the maxResults value provided in the request, enter the token returned in the nextToken field in the response in this field to return the next batch of results.
|
2358
|
+
*/
|
2359
|
+
nextToken?: PaginationToken;
|
2360
|
+
}
|
2361
|
+
export interface ListInferenceProfilesResponse {
|
2362
|
+
/**
|
2363
|
+
* A list of information about each inference profile that you can use.
|
2364
|
+
*/
|
2365
|
+
inferenceProfileSummaries?: InferenceProfileSummaries;
|
2366
|
+
/**
|
2367
|
+
* If the total number of results is greater than the maxResults value provided in the request, use this token when making another request in the nextToken field to return the next batch of results.
|
2368
|
+
*/
|
2369
|
+
nextToken?: PaginationToken;
|
2370
|
+
}
|
2238
2371
|
export interface ListModelCopyJobsRequest {
|
2239
2372
|
/**
|
2240
2373
|
* Filters for model copy jobs created after the specified time.
|