aws-sdk 2.1591.0 → 2.1592.0

Sign up to get free protection for your applications and to get access to all the features.
@@ -28,11 +28,11 @@ declare class CleanRoomsML extends Service {
28
28
  */
29
29
  createConfiguredAudienceModel(callback?: (err: AWSError, data: CleanRoomsML.Types.CreateConfiguredAudienceModelResponse) => void): Request<CleanRoomsML.Types.CreateConfiguredAudienceModelResponse, AWSError>;
30
30
  /**
31
- * Defines the information necessary to create a training dataset, or seed audience. In Clean Rooms ML, the TrainingDataset is metadata that points to a Glue table, which is read only during AudienceModel creation.
31
+ * Defines the information necessary to create a training dataset. In Clean Rooms ML, the TrainingDataset is metadata that points to a Glue table, which is read only during AudienceModel creation.
32
32
  */
33
33
  createTrainingDataset(params: CleanRoomsML.Types.CreateTrainingDatasetRequest, callback?: (err: AWSError, data: CleanRoomsML.Types.CreateTrainingDatasetResponse) => void): Request<CleanRoomsML.Types.CreateTrainingDatasetResponse, AWSError>;
34
34
  /**
35
- * Defines the information necessary to create a training dataset, or seed audience. In Clean Rooms ML, the TrainingDataset is metadata that points to a Glue table, which is read only during AudienceModel creation.
35
+ * Defines the information necessary to create a training dataset. In Clean Rooms ML, the TrainingDataset is metadata that points to a Glue table, which is read only during AudienceModel creation.
36
36
  */
37
37
  createTrainingDataset(callback?: (err: AWSError, data: CleanRoomsML.Types.CreateTrainingDatasetResponse) => void): Request<CleanRoomsML.Types.CreateTrainingDatasetResponse, AWSError>;
38
38
  /**
@@ -257,7 +257,7 @@ declare namespace CleanRoomsML {
257
257
  export type AudienceGenerationJobArn = string;
258
258
  export interface AudienceGenerationJobDataSource {
259
259
  /**
260
- * The Amazon S3 bucket where the training data for the configured audience is stored.
260
+ * Defines the Amazon S3 bucket where the seed audience for the generating audience is stored. A valid data source is a JSON line file in the following format: {"user_id": "111111"} {"user_id": "222222"} ...
261
261
  */
262
262
  dataSource: S3ConfigMap;
263
263
  /**
@@ -307,22 +307,6 @@ declare namespace CleanRoomsML {
307
307
  }
308
308
  export type AudienceModelArn = string;
309
309
  export type AudienceModelList = AudienceModelSummary[];
310
- export interface AudienceModelMetric {
311
- /**
312
- * The number of users that were used to generate these model metrics.
313
- */
314
- forTopKItemPredictions: Integer;
315
- /**
316
- * The audience model metric.
317
- */
318
- type: AudienceModelMetricType;
319
- /**
320
- * The value of the audience model metric
321
- */
322
- value: Double;
323
- }
324
- export type AudienceModelMetricType = "NORMALIZED_DISCOUNTED_CUMULATIVE_GAIN"|"MEAN_RECIPROCAL_RANK"|"PRECISION"|"RECALL"|string;
325
- export type AudienceModelMetrics = AudienceModelMetric[];
326
310
  export type AudienceModelStatus = "CREATE_PENDING"|"CREATE_IN_PROGRESS"|"CREATE_FAILED"|"ACTIVE"|"DELETE_PENDING"|"DELETE_IN_PROGRESS"|"DELETE_FAILED"|string;
327
311
  export interface AudienceModelSummary {
328
312
  /**
@@ -355,6 +339,10 @@ declare namespace CleanRoomsML {
355
339
  updateTime: SyntheticTimestamp_date_time;
356
340
  }
357
341
  export interface AudienceQualityMetrics {
342
+ /**
343
+ * The recall score of the generated audience. Recall is the percentage of the most similar users (by default, the most similar 20%) from a sample of the training data that are included in the seed audience by the audience generation job. Values range from 0-1, larger values indicate a better audience. A recall value approximately equal to the maximum bin size indicates that the audience model is equivalent to random selection.
344
+ */
345
+ recallMetric?: Double;
358
346
  /**
359
347
  * The relevance scores of the generated audience.
360
348
  */
@@ -455,7 +443,7 @@ declare namespace CleanRoomsML {
455
443
  */
456
444
  name: NameString;
457
445
  /**
458
- * The optional metadata that you apply to the resource to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. The following basic restrictions apply to tags: Maximum number of tags per resource - 50. For each resource, each tag key must be unique, and each tag key can have only one value. Maximum key length - 128 Unicode characters in UTF-8. Maximum value length - 256 Unicode characters in UTF-8. If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @. Tag keys and values are case sensitive. Do not use aws:, AWS:, or any upper or lowercase combination of such as a prefix for keys as it is reserved for AWS use. You cannot edit or delete tag keys with this prefix. Values can have this prefix. If a tag value has aws as its prefix but the key does not, then Forecast considers it to be a user tag and will count against the limit of 50 tags. Tags with only the key prefix of aws do not count against your tags per resource limit.
446
+ * The optional metadata that you apply to the resource to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. The following basic restrictions apply to tags: Maximum number of tags per resource - 50. For each resource, each tag key must be unique, and each tag key can have only one value. Maximum key length - 128 Unicode characters in UTF-8. Maximum value length - 256 Unicode characters in UTF-8. If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @. Tag keys and values are case sensitive. Do not use aws:, AWS:, or any upper or lowercase combination of such as a prefix for keys as it is reserved for AWS use. You cannot edit or delete tag keys with this prefix. Values can have this prefix. If a tag value has aws as its prefix but the key does not, then Clean Rooms ML considers it to be a user tag and will count against the limit of 50 tags. Tags with only the key prefix of aws do not count against your tags per resource limit.
459
447
  */
460
448
  tags?: TagMap;
461
449
  /**
@@ -495,7 +483,7 @@ declare namespace CleanRoomsML {
495
483
  */
496
484
  description?: ResourceDescription;
497
485
  /**
498
- * The minimum number of users from the seed audience that must match with users in the training data of the audience model.
486
+ * The minimum number of users from the seed audience that must match with users in the training data of the audience model. The default value is 500.
499
487
  */
500
488
  minMatchingSeedSize?: MinMatchingSeedSize;
501
489
  /**
@@ -511,7 +499,7 @@ declare namespace CleanRoomsML {
511
499
  */
512
500
  sharedAudienceMetrics: MetricsList;
513
501
  /**
514
- * The optional metadata that you apply to the resource to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. The following basic restrictions apply to tags: Maximum number of tags per resource - 50. For each resource, each tag key must be unique, and each tag key can have only one value. Maximum key length - 128 Unicode characters in UTF-8. Maximum value length - 256 Unicode characters in UTF-8. If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @. Tag keys and values are case sensitive. Do not use aws:, AWS:, or any upper or lowercase combination of such as a prefix for keys as it is reserved for AWS use. You cannot edit or delete tag keys with this prefix. Values can have this prefix. If a tag value has aws as its prefix but the key does not, then Forecast considers it to be a user tag and will count against the limit of 50 tags. Tags with only the key prefix of aws do not count against your tags per resource limit.
502
+ * The optional metadata that you apply to the resource to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. The following basic restrictions apply to tags: Maximum number of tags per resource - 50. For each resource, each tag key must be unique, and each tag key can have only one value. Maximum key length - 128 Unicode characters in UTF-8. Maximum value length - 256 Unicode characters in UTF-8. If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @. Tag keys and values are case sensitive. Do not use aws:, AWS:, or any upper or lowercase combination of such as a prefix for keys as it is reserved for AWS use. You cannot edit or delete tag keys with this prefix. Values can have this prefix. If a tag value has aws as its prefix but the key does not, then Clean Rooms ML considers it to be a user tag and will count against the limit of 50 tags. Tags with only the key prefix of aws do not count against your tags per resource limit.
515
503
  */
516
504
  tags?: TagMap;
517
505
  }
@@ -642,7 +630,7 @@ declare namespace CleanRoomsML {
642
630
  */
643
631
  includeSeedInOutput?: Boolean;
644
632
  /**
645
- * The relevance scores for different audience sizes.
633
+ * The relevance scores for different audience sizes and the recall score of the generated audience.
646
634
  */
647
635
  metrics?: AudienceQualityMetrics;
648
636
  /**
@@ -697,10 +685,6 @@ declare namespace CleanRoomsML {
697
685
  * The KMS key ARN used for the audience model.
698
686
  */
699
687
  kmsKeyArn?: KmsKeyArn;
700
- /**
701
- * Accuracy metrics for the model.
702
- */
703
- metrics?: AudienceModelMetrics;
704
688
  /**
705
689
  * The name of the audience model.
706
690
  */
@@ -876,7 +860,6 @@ declare namespace CleanRoomsML {
876
860
  export type GlueTableName = string;
877
861
  export type Hash = string;
878
862
  export type IamRoleArn = string;
879
- export type Integer = number;
880
863
  export type KmsKeyArn = string;
881
864
  export interface ListAudienceExportJobsRequest {
882
865
  /**
@@ -1095,7 +1078,7 @@ declare namespace CleanRoomsML {
1095
1078
  */
1096
1079
  seedAudience: AudienceGenerationJobDataSource;
1097
1080
  /**
1098
- * The optional metadata that you apply to the resource to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. The following basic restrictions apply to tags: Maximum number of tags per resource - 50. For each resource, each tag key must be unique, and each tag key can have only one value. Maximum key length - 128 Unicode characters in UTF-8. Maximum value length - 256 Unicode characters in UTF-8. If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @. Tag keys and values are case sensitive. Do not use aws:, AWS:, or any upper or lowercase combination of such as a prefix for keys as it is reserved for AWS use. You cannot edit or delete tag keys with this prefix. Values can have this prefix. If a tag value has aws as its prefix but the key does not, then Forecast considers it to be a user tag and will count against the limit of 50 tags. Tags with only the key prefix of aws do not count against your tags per resource limit.
1081
+ * The optional metadata that you apply to the resource to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. The following basic restrictions apply to tags: Maximum number of tags per resource - 50. For each resource, each tag key must be unique, and each tag key can have only one value. Maximum key length - 128 Unicode characters in UTF-8. Maximum value length - 256 Unicode characters in UTF-8. If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @. Tag keys and values are case sensitive. Do not use aws:, AWS:, or any upper or lowercase combination of such as a prefix for keys as it is reserved for AWS use. You cannot edit or delete tag keys with this prefix. Values can have this prefix. If a tag value has aws as its prefix but the key does not, then Clean Rooms ML considers it to be a user tag and will count against the limit of 50 tags. Tags with only the key prefix of aws do not count against your tags per resource limit.
1099
1082
  */
1100
1083
  tags?: TagMap;
1101
1084
  }
@@ -1127,7 +1110,7 @@ declare namespace CleanRoomsML {
1127
1110
  */
1128
1111
  resourceArn: TaggableArn;
1129
1112
  /**
1130
- * The optional metadata that you apply to the resource to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. The following basic restrictions apply to tags: Maximum number of tags per resource - 50. For each resource, each tag key must be unique, and each tag key can have only one value. Maximum key length - 128 Unicode characters in UTF-8. Maximum value length - 256 Unicode characters in UTF-8. If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @. Tag keys and values are case sensitive. Do not use aws:, AWS:, or any upper or lowercase combination of such as a prefix for keys as it is reserved for AWS use. You cannot edit or delete tag keys with this prefix. Values can have this prefix. If a tag value has aws as its prefix but the key does not, then Forecast considers it to be a user tag and will count against the limit of 50 tags. Tags with only the key prefix of aws do not count against your tags per resource limit.
1113
+ * The optional metadata that you apply to the resource to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. The following basic restrictions apply to tags: Maximum number of tags per resource - 50. For each resource, each tag key must be unique, and each tag key can have only one value. Maximum key length - 128 Unicode characters in UTF-8. Maximum value length - 256 Unicode characters in UTF-8. If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @. Tag keys and values are case sensitive. Do not use aws:, AWS:, or any upper or lowercase combination of such as a prefix for keys as it is reserved for AWS use. You cannot edit or delete tag keys with this prefix. Values can have this prefix. If a tag value has aws as its prefix but the key does not, then Clean Rooms considers it to be a user tag and will count against the limit of 50 tags. Tags with only the key prefix of aws do not count against your tags per resource limit.
1131
1114
  */
1132
1115
  tags: TagMap;
1133
1116
  }
@@ -3147,6 +3147,7 @@ declare namespace CloudFormation {
3147
3147
  */
3148
3148
  Value: Value;
3149
3149
  }
3150
+ export type PolicyAction = "Delete"|"Retain"|"Snapshot"|"ReplaceAndDelete"|"ReplaceAndRetain"|"ReplaceAndSnapshot"|string;
3150
3151
  export type PrivateTypeArn = string;
3151
3152
  export type Properties = string;
3152
3153
  export type PropertyDescription = string;
@@ -3317,6 +3318,10 @@ declare namespace CloudFormation {
3317
3318
  export type RequiresRecreation = "Never"|"Conditionally"|"Always"|string;
3318
3319
  export type ResourceAttribute = "Properties"|"Metadata"|"CreationPolicy"|"UpdatePolicy"|"DeletionPolicy"|"UpdateReplacePolicy"|"Tags"|string;
3319
3320
  export interface ResourceChange {
3321
+ /**
3322
+ * The action that will be taken on the physical resource when the change set is executed. Delete The resource will be deleted. Retain The resource will be retained. Snapshot The resource will have a snapshot taken. ReplaceAndDelete The resource will be replaced and then deleted. ReplaceAndRetain The resource will be replaced and then retained. ReplaceAndSnapshot The resource will be replaced and then have a snapshot taken.
3323
+ */
3324
+ PolicyAction?: PolicyAction;
3320
3325
  /**
3321
3326
  * The action that CloudFormation takes on the resource, such as Add (adds a new resource), Modify (changes a resource), Remove (deletes a resource), Import (imports a resource), or Dynamic (exact action for the resource can't be determined).
3322
3327
  */
@@ -315,6 +315,14 @@ declare class DataZone extends Service {
315
315
  * Deletes a subscription target in Amazon DataZone.
316
316
  */
317
317
  deleteSubscriptionTarget(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
318
+ /**
319
+ * Deletes the specified time series form for the specified asset.
320
+ */
321
+ deleteTimeSeriesDataPoints(params: DataZone.Types.DeleteTimeSeriesDataPointsInput, callback?: (err: AWSError, data: DataZone.Types.DeleteTimeSeriesDataPointsOutput) => void): Request<DataZone.Types.DeleteTimeSeriesDataPointsOutput, AWSError>;
322
+ /**
323
+ * Deletes the specified time series form for the specified asset.
324
+ */
325
+ deleteTimeSeriesDataPoints(callback?: (err: AWSError, data: DataZone.Types.DeleteTimeSeriesDataPointsOutput) => void): Request<DataZone.Types.DeleteTimeSeriesDataPointsOutput, AWSError>;
318
326
  /**
319
327
  * Gets an Amazon DataZone asset.
320
328
  */
@@ -483,6 +491,14 @@ declare class DataZone extends Service {
483
491
  * Gets the subscription target in Amazon DataZone.
484
492
  */
485
493
  getSubscriptionTarget(callback?: (err: AWSError, data: DataZone.Types.GetSubscriptionTargetOutput) => void): Request<DataZone.Types.GetSubscriptionTargetOutput, AWSError>;
494
+ /**
495
+ * Gets the existing data point for the asset.
496
+ */
497
+ getTimeSeriesDataPoint(params: DataZone.Types.GetTimeSeriesDataPointInput, callback?: (err: AWSError, data: DataZone.Types.GetTimeSeriesDataPointOutput) => void): Request<DataZone.Types.GetTimeSeriesDataPointOutput, AWSError>;
498
+ /**
499
+ * Gets the existing data point for the asset.
500
+ */
501
+ getTimeSeriesDataPoint(callback?: (err: AWSError, data: DataZone.Types.GetTimeSeriesDataPointOutput) => void): Request<DataZone.Types.GetTimeSeriesDataPointOutput, AWSError>;
486
502
  /**
487
503
  * Gets a user profile in Amazon DataZone.
488
504
  */
@@ -635,6 +651,22 @@ declare class DataZone extends Service {
635
651
  * Lists tags for the specified resource in Amazon DataZone.
636
652
  */
637
653
  listTagsForResource(callback?: (err: AWSError, data: DataZone.Types.ListTagsForResourceResponse) => void): Request<DataZone.Types.ListTagsForResourceResponse, AWSError>;
654
+ /**
655
+ * Lists time series data points.
656
+ */
657
+ listTimeSeriesDataPoints(params: DataZone.Types.ListTimeSeriesDataPointsInput, callback?: (err: AWSError, data: DataZone.Types.ListTimeSeriesDataPointsOutput) => void): Request<DataZone.Types.ListTimeSeriesDataPointsOutput, AWSError>;
658
+ /**
659
+ * Lists time series data points.
660
+ */
661
+ listTimeSeriesDataPoints(callback?: (err: AWSError, data: DataZone.Types.ListTimeSeriesDataPointsOutput) => void): Request<DataZone.Types.ListTimeSeriesDataPointsOutput, AWSError>;
662
+ /**
663
+ * Posts time series data points to Amazon DataZone for the specified asset.
664
+ */
665
+ postTimeSeriesDataPoints(params: DataZone.Types.PostTimeSeriesDataPointsInput, callback?: (err: AWSError, data: DataZone.Types.PostTimeSeriesDataPointsOutput) => void): Request<DataZone.Types.PostTimeSeriesDataPointsOutput, AWSError>;
666
+ /**
667
+ * Posts time series data points to Amazon DataZone for the specified asset.
668
+ */
669
+ postTimeSeriesDataPoints(callback?: (err: AWSError, data: DataZone.Types.PostTimeSeriesDataPointsOutput) => void): Request<DataZone.Types.PostTimeSeriesDataPointsOutput, AWSError>;
638
670
  /**
639
671
  * Writes the configuration for the specified environment blueprint in Amazon DataZone.
640
672
  */
@@ -1036,6 +1068,10 @@ declare namespace DataZone {
1036
1068
  * The forms included in the additional attributes of an inventory asset.
1037
1069
  */
1038
1070
  formsOutput?: FormOutputList;
1071
+ /**
1072
+ * The latest time series data points forms included in the additional attributes of an asset.
1073
+ */
1074
+ latestTimeSeriesDataPointFormsOutput?: TimeSeriesDataPointSummaryFormOutputList;
1039
1075
  /**
1040
1076
  * The read-only forms included in the additional attributes of an inventory asset.
1041
1077
  */
@@ -1066,6 +1102,10 @@ declare namespace DataZone {
1066
1102
  * The glossary terms attached to an asset published in an Amazon DataZone catalog.
1067
1103
  */
1068
1104
  glossaryTerms?: DetailedGlossaryTerms;
1105
+ /**
1106
+ * The latest time series data points forms included in the additional attributes of an asset.
1107
+ */
1108
+ latestTimeSeriesDataPointForms?: TimeSeriesDataPointSummaryFormOutputList;
1069
1109
  /**
1070
1110
  * The identifier of the project where an asset published in an Amazon DataZone catalog exists.
1071
1111
  */
@@ -1140,6 +1180,10 @@ declare namespace DataZone {
1140
1180
  * The metadata forms that form additional attributes of the metadata asset.
1141
1181
  */
1142
1182
  forms?: Forms;
1183
+ /**
1184
+ * The latest time series data points forms included in the additional attributes of an asset.
1185
+ */
1186
+ latestTimeSeriesDataPointForms?: TimeSeriesDataPointSummaryFormOutputList;
1143
1187
  }
1144
1188
  export type AssetName = string;
1145
1189
  export interface AssetRevision {
@@ -1429,6 +1473,10 @@ declare namespace DataZone {
1429
1473
  * The unique identifier of the created asset.
1430
1474
  */
1431
1475
  id: AssetId;
1476
+ /**
1477
+ * The latest data point that was imported into the time series form for the asset.
1478
+ */
1479
+ latestTimeSeriesDataPointFormsOutput?: TimeSeriesDataPointSummaryFormOutputList;
1432
1480
  /**
1433
1481
  * The details of an asset published in an Amazon DataZone catalog.
1434
1482
  */
@@ -1541,6 +1589,10 @@ declare namespace DataZone {
1541
1589
  * The unique identifier of the asset revision.
1542
1590
  */
1543
1591
  id: AssetId;
1592
+ /**
1593
+ * The latest data point that was imported into the time series form for the asset.
1594
+ */
1595
+ latestTimeSeriesDataPointFormsOutput?: TimeSeriesDataPointSummaryFormOutputList;
1544
1596
  /**
1545
1597
  * The details of an asset published in an Amazon DataZone catalog.
1546
1598
  */
@@ -2278,7 +2330,7 @@ declare namespace DataZone {
2278
2330
  /**
2279
2331
  * The ID of the asset.
2280
2332
  */
2281
- entityIdentifier: EntityId;
2333
+ entityIdentifier: EntityIdentifier;
2282
2334
  /**
2283
2335
  * The revision of an asset.
2284
2336
  */
@@ -2701,6 +2753,7 @@ declare namespace DataZone {
2701
2753
  export type CustomParameterKeyNameString = string;
2702
2754
  export type CustomParameterList = CustomParameter[];
2703
2755
  export type DataAssetActivityStatus = "FAILED"|"PUBLISHING_FAILED"|"SUCCEEDED_CREATED"|"SUCCEEDED_UPDATED"|"SKIPPED_ALREADY_IMPORTED"|"SKIPPED_ARCHIVED"|"SKIPPED_NO_ACCESS"|"UNCHANGED"|string;
2756
+ export type DataPointIdentifier = string;
2704
2757
  export type DataProductDescription = string;
2705
2758
  export type DataProductId = string;
2706
2759
  export interface DataProductItem {
@@ -3262,6 +3315,30 @@ declare namespace DataZone {
3262
3315
  */
3263
3316
  identifier: SubscriptionTargetId;
3264
3317
  }
3318
+ export interface DeleteTimeSeriesDataPointsInput {
3319
+ /**
3320
+ * A unique, case-sensitive identifier to ensure idempotency of the request. This field is automatically populated if not provided.
3321
+ */
3322
+ clientToken?: ClientToken;
3323
+ /**
3324
+ * The ID of the Amazon DataZone domain that houses the asset for which you want to delete a time series form.
3325
+ */
3326
+ domainIdentifier: DomainId;
3327
+ /**
3328
+ * The ID of the asset for which you want to delete a time series form.
3329
+ */
3330
+ entityIdentifier: EntityIdentifier;
3331
+ /**
3332
+ * The type of the asset for which you want to delete a time series form.
3333
+ */
3334
+ entityType: TimeSeriesEntityType;
3335
+ /**
3336
+ * The name of the time series form that you want to delete.
3337
+ */
3338
+ formName: TimeSeriesFormName;
3339
+ }
3340
+ export interface DeleteTimeSeriesDataPointsOutput {
3341
+ }
3265
3342
  export interface Deployment {
3266
3343
  /**
3267
3344
  * The identifier of the last deployment of the environment.
@@ -3363,6 +3440,7 @@ declare namespace DataZone {
3363
3440
  export type EnableSetting = "ENABLED"|"DISABLED"|string;
3364
3441
  export type EnabledRegionList = RegionName[];
3365
3442
  export type EntityId = string;
3443
+ export type EntityIdentifier = string;
3366
3444
  export type EntityType = "ASSET"|string;
3367
3445
  export type EnvironmentActionList = ConfigurableEnvironmentAction[];
3368
3446
  export interface EnvironmentBlueprintConfigurationItem {
@@ -3789,6 +3867,10 @@ declare namespace DataZone {
3789
3867
  * The ID of the asset.
3790
3868
  */
3791
3869
  id: AssetId;
3870
+ /**
3871
+ * The latest data point that was imported into the time series form for the asset.
3872
+ */
3873
+ latestTimeSeriesDataPointFormsOutput?: TimeSeriesDataPointSummaryFormOutputList;
3792
3874
  /**
3793
3875
  * The listing of the asset.
3794
3876
  */
@@ -4976,6 +5058,50 @@ declare namespace DataZone {
4976
5058
  */
4977
5059
  updatedBy?: UpdatedBy;
4978
5060
  }
5061
+ export interface GetTimeSeriesDataPointInput {
5062
+ /**
5063
+ * The ID of the Amazon DataZone domain that houses the asset for which you want to get the data point.
5064
+ */
5065
+ domainIdentifier: DomainId;
5066
+ /**
5067
+ * The ID of the asset for which you want to get the data point.
5068
+ */
5069
+ entityIdentifier: EntityIdentifier;
5070
+ /**
5071
+ * The type of the asset for which you want to get the data point.
5072
+ */
5073
+ entityType: TimeSeriesEntityType;
5074
+ /**
5075
+ * The name of the time series form that houses the data point that you want to get.
5076
+ */
5077
+ formName: TimeSeriesFormName;
5078
+ /**
5079
+ * The ID of the data point that you want to get.
5080
+ */
5081
+ identifier: TimeSeriesDataPointIdentifier;
5082
+ }
5083
+ export interface GetTimeSeriesDataPointOutput {
5084
+ /**
5085
+ * The ID of the Amazon DataZone domain that houses the asset data point that you want to get.
5086
+ */
5087
+ domainId?: DomainId;
5088
+ /**
5089
+ * The ID of the asset for which you want to get the data point.
5090
+ */
5091
+ entityId?: EntityId;
5092
+ /**
5093
+ * The type of the asset for which you want to get the data point.
5094
+ */
5095
+ entityType?: TimeSeriesEntityType;
5096
+ /**
5097
+ * The time series form that houses the data point that you want to get.
5098
+ */
5099
+ form?: TimeSeriesDataPointFormOutput;
5100
+ /**
5101
+ * The name of the time series form that houses the data point that you want to get.
5102
+ */
5103
+ formName?: TimeSeriesFormName;
5104
+ }
4979
5105
  export interface GetUserProfileInput {
4980
5106
  /**
4981
5107
  * the ID of the Amazon DataZone domain the data portal of which you want to get.
@@ -5110,6 +5236,10 @@ declare namespace DataZone {
5110
5236
  export type GlossaryTermStatus = "ENABLED"|"DISABLED"|string;
5111
5237
  export type GlossaryTerms = GlossaryTermId[];
5112
5238
  export interface GlueRunConfigurationInput {
5239
+ /**
5240
+ * Specifies whether to automatically import data quality metrics as part of the data source run.
5241
+ */
5242
+ autoImportDataQualityResult?: Boolean;
5113
5243
  /**
5114
5244
  * The data access role included in the configuration details of the Amazon Web Services Glue data source.
5115
5245
  */
@@ -5125,6 +5255,10 @@ declare namespace DataZone {
5125
5255
  * The Amazon Web Services account ID included in the configuration details of the Amazon Web Services Glue data source.
5126
5256
  */
5127
5257
  accountId?: GlueRunConfigurationOutputAccountIdString;
5258
+ /**
5259
+ * Specifies whether to automatically import data quality metrics as part of the data source run.
5260
+ */
5261
+ autoImportDataQualityResult?: Boolean;
5128
5262
  /**
5129
5263
  * The data access role included in the configuration details of the Amazon Web Services Glue data source.
5130
5264
  */
@@ -5865,6 +5999,50 @@ declare namespace DataZone {
5865
5999
  */
5866
6000
  tags?: Tags;
5867
6001
  }
6002
+ export interface ListTimeSeriesDataPointsInput {
6003
+ /**
6004
+ * The ID of the Amazon DataZone domain that houses the assets for which you want to list time series data points.
6005
+ */
6006
+ domainIdentifier: DomainId;
6007
+ /**
6008
+ * The timestamp at which the data points that you wanted to list ended.
6009
+ */
6010
+ endedAt?: Timestamp;
6011
+ /**
6012
+ * The ID of the asset for which you want to list data points.
6013
+ */
6014
+ entityIdentifier: EntityIdentifier;
6015
+ /**
6016
+ * The type of the asset for which you want to list data points.
6017
+ */
6018
+ entityType: TimeSeriesEntityType;
6019
+ /**
6020
+ * The name of the time series data points form.
6021
+ */
6022
+ formName: TimeSeriesFormName;
6023
+ /**
6024
+ * The maximum number of data points to return in a single call to ListTimeSeriesDataPoints. When the number of data points to be listed is greater than the value of MaxResults, the response contains a NextToken value that you can use in a subsequent call to ListTimeSeriesDataPoints to list the next set of data points.
6025
+ */
6026
+ maxResults?: MaxResults;
6027
+ /**
6028
+ * When the number of data points is greater than the default value for the MaxResults parameter, or if you explicitly specify a value for MaxResults that is less than the number of data points, the response includes a pagination token named NextToken. You can specify this NextToken value in a subsequent call to ListTimeSeriesDataPoints to list the next set of data points.
6029
+ */
6030
+ nextToken?: PaginationToken;
6031
+ /**
6032
+ * The timestamp at which the data points that you want to list started.
6033
+ */
6034
+ startedAt?: Timestamp;
6035
+ }
6036
+ export interface ListTimeSeriesDataPointsOutput {
6037
+ /**
6038
+ * The results of the ListTimeSeriesDataPoints action.
6039
+ */
6040
+ items?: TimeSeriesDataPointSummaryFormOutputList;
6041
+ /**
6042
+ * When the number of data points is greater than the default value for the MaxResults parameter, or if you explicitly specify a value for MaxResults that is less than the number of data points, the response includes a pagination token named NextToken. You can specify this NextToken value in a subsequent call to ListTimeSeriesDataPoints to list the next set of data points.
6043
+ */
6044
+ nextToken?: PaginationToken;
6045
+ }
5868
6046
  export type ListingId = string;
5869
6047
  export interface ListingItem {
5870
6048
  /**
@@ -6045,6 +6223,46 @@ declare namespace DataZone {
6045
6223
  export type NotificationType = "TASK"|"EVENT"|string;
6046
6224
  export type NotificationsList = NotificationOutput[];
6047
6225
  export type PaginationToken = string;
6226
+ export interface PostTimeSeriesDataPointsInput {
6227
+ /**
6228
+ * A unique, case-sensitive identifier that is provided to ensure the idempotency of the request.
6229
+ */
6230
+ clientToken?: ClientToken;
6231
+ /**
6232
+ * The ID of the Amazon DataZone domain in which you want to post time series data points.
6233
+ */
6234
+ domainIdentifier: DomainId;
6235
+ /**
6236
+ * The ID of the asset for which you want to post time series data points.
6237
+ */
6238
+ entityIdentifier: EntityIdentifier;
6239
+ /**
6240
+ * The type of the asset for which you want to post data points.
6241
+ */
6242
+ entityType: TimeSeriesEntityType;
6243
+ /**
6244
+ * The forms that contain the data points that you want to post.
6245
+ */
6246
+ forms: TimeSeriesDataPointFormInputList;
6247
+ }
6248
+ export interface PostTimeSeriesDataPointsOutput {
6249
+ /**
6250
+ * The ID of the Amazon DataZone domain in which you want to post time series data points.
6251
+ */
6252
+ domainId?: DomainId;
6253
+ /**
6254
+ * The ID of the asset for which you want to post time series data points.
6255
+ */
6256
+ entityId?: EntityId;
6257
+ /**
6258
+ * The type of the asset for which you want to post data points.
6259
+ */
6260
+ entityType?: TimeSeriesEntityType;
6261
+ /**
6262
+ * The forms that contain the data points that you have posted.
6263
+ */
6264
+ forms?: TimeSeriesDataPointFormOutputList;
6265
+ }
6048
6266
  export type PredictionChoices = Integer[];
6049
6267
  export interface PredictionConfiguration {
6050
6268
  /**
@@ -6678,7 +6896,7 @@ declare namespace DataZone {
6678
6896
  */
6679
6897
  totalMatchCount?: Integer;
6680
6898
  }
6681
- export type SearchOutputAdditionalAttribute = "FORMS"|string;
6899
+ export type SearchOutputAdditionalAttribute = "FORMS"|"TIME_SERIES_DATA_POINT_FORMS"|string;
6682
6900
  export type SearchOutputAdditionalAttributes = SearchOutputAdditionalAttribute[];
6683
6901
  export interface SearchResultItem {
6684
6902
  /**
@@ -7329,6 +7547,89 @@ declare namespace DataZone {
7329
7547
  }
7330
7548
  export type TermRelationsClassifiesList = GlossaryTermId[];
7331
7549
  export type TermRelationsIsAList = GlossaryTermId[];
7550
+ export interface TimeSeriesDataPointFormInput {
7551
+ /**
7552
+ * The content of the time series data points form.
7553
+ */
7554
+ content?: TimeSeriesDataPointFormInputContentString;
7555
+ /**
7556
+ * The name of the time series data points form.
7557
+ */
7558
+ formName: TimeSeriesFormName;
7559
+ /**
7560
+ * The timestamp of the time series data points form.
7561
+ */
7562
+ timestamp: Timestamp;
7563
+ /**
7564
+ * The ID of the type of the time series data points form.
7565
+ */
7566
+ typeIdentifier: FormTypeIdentifier;
7567
+ /**
7568
+ * The revision type of the time series data points form.
7569
+ */
7570
+ typeRevision?: Revision;
7571
+ }
7572
+ export type TimeSeriesDataPointFormInputContentString = string;
7573
+ export type TimeSeriesDataPointFormInputList = TimeSeriesDataPointFormInput[];
7574
+ export interface TimeSeriesDataPointFormOutput {
7575
+ /**
7576
+ * The content of the time series data points form.
7577
+ */
7578
+ content?: TimeSeriesDataPointFormOutputContentString;
7579
+ /**
7580
+ * The name of the time series data points form.
7581
+ */
7582
+ formName: TimeSeriesFormName;
7583
+ /**
7584
+ * The ID of the time series data points form.
7585
+ */
7586
+ id?: DataPointIdentifier;
7587
+ /**
7588
+ * The timestamp of the time series data points form.
7589
+ */
7590
+ timestamp: Timestamp;
7591
+ /**
7592
+ * The ID of the type of the time series data points form.
7593
+ */
7594
+ typeIdentifier: FormTypeIdentifier;
7595
+ /**
7596
+ * The revision type of the time series data points form.
7597
+ */
7598
+ typeRevision?: Revision;
7599
+ }
7600
+ export type TimeSeriesDataPointFormOutputContentString = string;
7601
+ export type TimeSeriesDataPointFormOutputList = TimeSeriesDataPointFormOutput[];
7602
+ export type TimeSeriesDataPointIdentifier = string;
7603
+ export interface TimeSeriesDataPointSummaryFormOutput {
7604
+ /**
7605
+ * The content of the summary of the time series data points form.
7606
+ */
7607
+ contentSummary?: TimeSeriesDataPointSummaryFormOutputContentSummaryString;
7608
+ /**
7609
+ * The name of the time series data points summary form.
7610
+ */
7611
+ formName: TimeSeriesFormName;
7612
+ /**
7613
+ * The ID of the time series data points summary form.
7614
+ */
7615
+ id?: DataPointIdentifier;
7616
+ /**
7617
+ * The timestamp of the time series data points summary form.
7618
+ */
7619
+ timestamp: Timestamp;
7620
+ /**
7621
+ * The type ID of the time series data points summary form.
7622
+ */
7623
+ typeIdentifier: FormTypeIdentifier;
7624
+ /**
7625
+ * The type revision of the time series data points summary form.
7626
+ */
7627
+ typeRevision?: Revision;
7628
+ }
7629
+ export type TimeSeriesDataPointSummaryFormOutputContentSummaryString = string;
7630
+ export type TimeSeriesDataPointSummaryFormOutputList = TimeSeriesDataPointSummaryFormOutput[];
7631
+ export type TimeSeriesEntityType = "ASSET"|"LISTING"|string;
7632
+ export type TimeSeriesFormName = string;
7332
7633
  export type Timestamp = Date;
7333
7634
  export type Timezone = "UTC"|"AFRICA_JOHANNESBURG"|"AMERICA_MONTREAL"|"AMERICA_SAO_PAULO"|"ASIA_BAHRAIN"|"ASIA_BANGKOK"|"ASIA_CALCUTTA"|"ASIA_DUBAI"|"ASIA_HONG_KONG"|"ASIA_JAKARTA"|"ASIA_KUALA_LUMPUR"|"ASIA_SEOUL"|"ASIA_SHANGHAI"|"ASIA_SINGAPORE"|"ASIA_TAIPEI"|"ASIA_TOKYO"|"AUSTRALIA_MELBOURNE"|"AUSTRALIA_SYDNEY"|"CANADA_CENTRAL"|"CET"|"CST6CDT"|"ETC_GMT"|"ETC_GMT0"|"ETC_GMT_ADD_0"|"ETC_GMT_ADD_1"|"ETC_GMT_ADD_10"|"ETC_GMT_ADD_11"|"ETC_GMT_ADD_12"|"ETC_GMT_ADD_2"|"ETC_GMT_ADD_3"|"ETC_GMT_ADD_4"|"ETC_GMT_ADD_5"|"ETC_GMT_ADD_6"|"ETC_GMT_ADD_7"|"ETC_GMT_ADD_8"|"ETC_GMT_ADD_9"|"ETC_GMT_NEG_0"|"ETC_GMT_NEG_1"|"ETC_GMT_NEG_10"|"ETC_GMT_NEG_11"|"ETC_GMT_NEG_12"|"ETC_GMT_NEG_13"|"ETC_GMT_NEG_14"|"ETC_GMT_NEG_2"|"ETC_GMT_NEG_3"|"ETC_GMT_NEG_4"|"ETC_GMT_NEG_5"|"ETC_GMT_NEG_6"|"ETC_GMT_NEG_7"|"ETC_GMT_NEG_8"|"ETC_GMT_NEG_9"|"EUROPE_DUBLIN"|"EUROPE_LONDON"|"EUROPE_PARIS"|"EUROPE_STOCKHOLM"|"EUROPE_ZURICH"|"ISRAEL"|"MEXICO_GENERAL"|"MST7MDT"|"PACIFIC_AUCKLAND"|"US_CENTRAL"|"US_EASTERN"|"US_MOUNTAIN"|"US_PACIFIC"|string;
7334
7635
  export type Title = string;
@@ -436,6 +436,14 @@ declare class DocDB extends Service {
436
436
  * Stops the running cluster that is specified by DBClusterIdentifier. The cluster must be in the available state. For more information, see Stopping and Starting an Amazon DocumentDB Cluster.
437
437
  */
438
438
  stopDBCluster(callback?: (err: AWSError, data: DocDB.Types.StopDBClusterResult) => void): Request<DocDB.Types.StopDBClusterResult, AWSError>;
439
+ /**
440
+ * Switches over the specified secondary Amazon DocumentDB cluster to be the new primary Amazon DocumentDB cluster in the global database cluster.
441
+ */
442
+ switchoverGlobalCluster(params: DocDB.Types.SwitchoverGlobalClusterMessage, callback?: (err: AWSError, data: DocDB.Types.SwitchoverGlobalClusterResult) => void): Request<DocDB.Types.SwitchoverGlobalClusterResult, AWSError>;
443
+ /**
444
+ * Switches over the specified secondary Amazon DocumentDB cluster to be the new primary Amazon DocumentDB cluster in the global database cluster.
445
+ */
446
+ switchoverGlobalCluster(callback?: (err: AWSError, data: DocDB.Types.SwitchoverGlobalClusterResult) => void): Request<DocDB.Types.SwitchoverGlobalClusterResult, AWSError>;
439
447
  /**
440
448
  * Waits for the dBInstanceAvailable state by periodically calling the underlying DocDB.describeDBInstancesoperation every 30 seconds (at most 60 times).
441
449
  */
@@ -2740,6 +2748,19 @@ declare namespace DocDB {
2740
2748
  }
2741
2749
  export type SubnetIdentifierList = String[];
2742
2750
  export type SubnetList = Subnet[];
2751
+ export interface SwitchoverGlobalClusterMessage {
2752
+ /**
2753
+ * The identifier of the Amazon DocumentDB global database cluster to switch over. The identifier is the unique key assigned by the user when the cluster is created. In other words, it's the name of the global cluster. This parameter isn’t case-sensitive. Constraints: Must match the identifier of an existing global cluster (Amazon DocumentDB global database). Minimum length of 1. Maximum length of 255. Pattern: [A-Za-z][0-9A-Za-z-:._]*
2754
+ */
2755
+ GlobalClusterIdentifier: GlobalClusterIdentifier;
2756
+ /**
2757
+ * The identifier of the secondary Amazon DocumentDB cluster to promote to the new primary for the global database cluster. Use the Amazon Resource Name (ARN) for the identifier so that Amazon DocumentDB can locate the cluster in its Amazon Web Services region. Constraints: Must match the identifier of an existing secondary cluster. Minimum length of 1. Maximum length of 255. Pattern: [A-Za-z][0-9A-Za-z-:._]*
2758
+ */
2759
+ TargetDbClusterIdentifier: String;
2760
+ }
2761
+ export interface SwitchoverGlobalClusterResult {
2762
+ GlobalCluster?: GlobalCluster;
2763
+ }
2743
2764
  export type TStamp = Date;
2744
2765
  export interface Tag {
2745
2766
  /**