@aws-sdk/client-personalize 3.451.0 → 3.458.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,8 +1,12 @@
1
1
  "use strict";
2
2
  Object.defineProperty(exports, "__esModule", { value: true });
3
- exports.DescribeFilterResponseFilterSensitiveLog = exports.FilterFilterSensitiveLog = exports.CreateFilterRequestFilterSensitiveLog = exports.TooManyTagKeysException = exports.RecipeProvider = exports.InvalidNextTokenException = exports.TrainingMode = exports.ObjectiveSensitivity = exports.ImportMode = exports.Domain = exports.IngestionMode = exports.TooManyTagsException = exports.ResourceNotFoundException = exports.ResourceInUseException = exports.ResourceAlreadyExistsException = exports.LimitExceededException = exports.InvalidInputException = void 0;
3
+ exports.DescribeFilterResponseFilterSensitiveLog = exports.FilterFilterSensitiveLog = exports.CreateFilterRequestFilterSensitiveLog = exports.TooManyTagKeysException = exports.RecipeProvider = exports.InvalidNextTokenException = exports.TrainingMode = exports.ObjectiveSensitivity = exports.ImportMode = exports.Domain = exports.IngestionMode = exports.TooManyTagsException = exports.ResourceNotFoundException = exports.ResourceInUseException = exports.ResourceAlreadyExistsException = exports.LimitExceededException = exports.InvalidInputException = exports.BatchInferenceJobMode = void 0;
4
4
  const smithy_client_1 = require("@smithy/smithy-client");
5
5
  const PersonalizeServiceException_1 = require("./PersonalizeServiceException");
6
+ exports.BatchInferenceJobMode = {
7
+ BATCH_INFERENCE: "BATCH_INFERENCE",
8
+ THEME_GENERATION: "THEME_GENERATION",
9
+ };
6
10
  class InvalidInputException extends PersonalizeServiceException_1.PersonalizeServiceException {
7
11
  constructor(opts) {
8
12
  super({
@@ -3215,6 +3215,7 @@ const de_BatchInferenceJob = (output, context) => {
3215
3215
  return (0, smithy_client_1.take)(output, {
3216
3216
  batchInferenceJobArn: smithy_client_1.expectString,
3217
3217
  batchInferenceJobConfig: smithy_client_1._json,
3218
+ batchInferenceJobMode: smithy_client_1.expectString,
3218
3219
  creationDateTime: (_) => (0, smithy_client_1.expectNonNull)((0, smithy_client_1.parseEpochTimestamp)((0, smithy_client_1.expectNumber)(_))),
3219
3220
  failureReason: smithy_client_1.expectString,
3220
3221
  filterArn: smithy_client_1.expectString,
@@ -3226,6 +3227,7 @@ const de_BatchInferenceJob = (output, context) => {
3226
3227
  roleArn: smithy_client_1.expectString,
3227
3228
  solutionVersionArn: smithy_client_1.expectString,
3228
3229
  status: smithy_client_1.expectString,
3230
+ themeGenerationConfig: smithy_client_1._json,
3229
3231
  });
3230
3232
  };
3231
3233
  const de_BatchInferenceJobs = (output, context) => {
@@ -3239,6 +3241,7 @@ const de_BatchInferenceJobs = (output, context) => {
3239
3241
  const de_BatchInferenceJobSummary = (output, context) => {
3240
3242
  return (0, smithy_client_1.take)(output, {
3241
3243
  batchInferenceJobArn: smithy_client_1.expectString,
3244
+ batchInferenceJobMode: smithy_client_1.expectString,
3242
3245
  creationDateTime: (_) => (0, smithy_client_1.expectNonNull)((0, smithy_client_1.parseEpochTimestamp)((0, smithy_client_1.expectNumber)(_))),
3243
3246
  failureReason: smithy_client_1.expectString,
3244
3247
  jobName: smithy_client_1.expectString,
@@ -3351,6 +3354,7 @@ const de_Dataset = (output, context) => {
3351
3354
  name: smithy_client_1.expectString,
3352
3355
  schemaArn: smithy_client_1.expectString,
3353
3356
  status: smithy_client_1.expectString,
3357
+ trackingId: smithy_client_1.expectString,
3354
3358
  });
3355
3359
  };
3356
3360
  const de_DatasetExportJob = (output, context) => {
@@ -1,5 +1,9 @@
1
1
  import { SENSITIVE_STRING } from "@smithy/smithy-client";
2
2
  import { PersonalizeServiceException as __BaseException } from "./PersonalizeServiceException";
3
+ export const BatchInferenceJobMode = {
4
+ BATCH_INFERENCE: "BATCH_INFERENCE",
5
+ THEME_GENERATION: "THEME_GENERATION",
6
+ };
3
7
  export class InvalidInputException extends __BaseException {
4
8
  constructor(opts) {
5
9
  super({
@@ -3076,6 +3076,7 @@ const de_BatchInferenceJob = (output, context) => {
3076
3076
  return take(output, {
3077
3077
  batchInferenceJobArn: __expectString,
3078
3078
  batchInferenceJobConfig: _json,
3079
+ batchInferenceJobMode: __expectString,
3079
3080
  creationDateTime: (_) => __expectNonNull(__parseEpochTimestamp(__expectNumber(_))),
3080
3081
  failureReason: __expectString,
3081
3082
  filterArn: __expectString,
@@ -3087,6 +3088,7 @@ const de_BatchInferenceJob = (output, context) => {
3087
3088
  roleArn: __expectString,
3088
3089
  solutionVersionArn: __expectString,
3089
3090
  status: __expectString,
3091
+ themeGenerationConfig: _json,
3090
3092
  });
3091
3093
  };
3092
3094
  const de_BatchInferenceJobs = (output, context) => {
@@ -3100,6 +3102,7 @@ const de_BatchInferenceJobs = (output, context) => {
3100
3102
  const de_BatchInferenceJobSummary = (output, context) => {
3101
3103
  return take(output, {
3102
3104
  batchInferenceJobArn: __expectString,
3105
+ batchInferenceJobMode: __expectString,
3103
3106
  creationDateTime: (_) => __expectNonNull(__parseEpochTimestamp(__expectNumber(_))),
3104
3107
  failureReason: __expectString,
3105
3108
  jobName: __expectString,
@@ -3212,6 +3215,7 @@ const de_Dataset = (output, context) => {
3212
3215
  name: __expectString,
3213
3216
  schemaArn: __expectString,
3214
3217
  status: __expectString,
3218
+ trackingId: __expectString,
3215
3219
  });
3216
3220
  };
3217
3221
  const de_DatasetExportJob = (output, context) => {
@@ -23,10 +23,24 @@ export interface CreateBatchInferenceJobCommandOutput extends CreateBatchInferen
23
23
  }
24
24
  /**
25
25
  * @public
26
- * <p>Creates a batch inference job. The operation can handle up to 50 million records and the
27
- * input file must be in JSON format. For more information, see
28
- * <a href="https://docs.aws.amazon.com/personalize/latest/dg/creating-batch-inference-job.html">Creating a batch inference job</a>.
26
+ * <p>Generates batch recommendations based on a list of items or users stored in Amazon S3
27
+ * and exports the recommendations to an Amazon S3 bucket.</p>
28
+ * <p>To generate batch recommendations, specify the ARN of a solution version and an Amazon S3 URI for the input and output data.
29
+ * For user personalization, popular items, and personalized ranking solutions, the batch inference job generates a list of
30
+ * recommended items for each user ID in the input file. For related items solutions, the job generates a list of recommended
31
+ * items for each item ID in the input file.</p>
32
+ * <p>For more information, see <a href="https://docs.aws.amazon.com/personalize/latest/dg/getting-batch-recommendations.html">Creating a batch inference job
33
+ * </a>.</p>
34
+ * <p>
35
+ * If you use the Similar-Items recipe, Amazon Personalize can add descriptive themes to batch recommendations.
36
+ * To generate themes, set the job's mode to
37
+ * <code>THEME_GENERATION</code> and specify the name of the field that contains item names in the
38
+ * input data.</p>
39
+ * <p>
40
+ * For more information about generating themes, see <a href="https://docs.aws.amazon.com/personalize/latest/dg/themed-batch-recommendations.html">Batch recommendations with themes from Content Generator
41
+ * </a>.
29
42
  * </p>
43
+ * <p>You can't get batch recommendations with the Trending-Now or Next-Best-Action recipes.</p>
30
44
  * @example
31
45
  * Use a bare-bones client and the command you need to make an API call.
32
46
  * ```javascript
@@ -62,6 +76,12 @@ export interface CreateBatchInferenceJobCommandOutput extends CreateBatchInferen
62
76
  * tagValue: "STRING_VALUE", // required
63
77
  * },
64
78
  * ],
79
+ * batchInferenceJobMode: "BATCH_INFERENCE" || "THEME_GENERATION",
80
+ * themeGenerationConfig: { // ThemeGenerationConfig
81
+ * fieldsForThemeGeneration: { // FieldsForThemeGeneration
82
+ * itemName: "STRING_VALUE", // required
83
+ * },
84
+ * },
65
85
  * };
66
86
  * const command = new CreateBatchInferenceJobCommand(input);
67
87
  * const response = await client.send(command);
@@ -37,22 +37,22 @@ export interface CreateCampaignCommandOutput extends CreateCampaignResponse, __M
37
37
  * as necessary.</p>
38
38
  * </important>
39
39
  * <p>A transaction is a single <code>GetRecommendations</code> or
40
- * <code>GetPersonalizedRanking</code> call. Transactions per second (TPS) is the throughput
41
- * and unit of billing for Amazon Personalize. The minimum provisioned TPS
42
- * (<code>minProvisionedTPS</code>) specifies the baseline throughput provisioned by
43
- * Amazon Personalize, and thus, the minimum billing charge.
40
+ * <code>GetPersonalizedRanking</code> call. Transactions per second (TPS) is the throughput
41
+ * and unit of billing for Amazon Personalize. The minimum provisioned TPS
42
+ * (<code>minProvisionedTPS</code>) specifies the baseline throughput provisioned by
43
+ * Amazon Personalize, and thus, the minimum billing charge.
44
44
  * </p>
45
45
  * <p>
46
- * If your TPS increases beyond
47
- * <code>minProvisionedTPS</code>, Amazon Personalize auto-scales the provisioned capacity up and down,
48
- * but never below <code>minProvisionedTPS</code>.
49
- * There's a short time delay while the capacity is increased that might cause loss of
50
- * transactions.</p>
46
+ * If your TPS increases beyond
47
+ * <code>minProvisionedTPS</code>, Amazon Personalize auto-scales the provisioned capacity up and down,
48
+ * but never below <code>minProvisionedTPS</code>.
49
+ * There's a short time delay while the capacity is increased that might cause loss of
50
+ * transactions.</p>
51
51
  * <p>The actual TPS used is calculated as the average requests/second within a 5-minute window.
52
52
  * You pay for maximum of either the minimum provisioned TPS or the actual TPS.
53
53
  * We recommend starting with a low <code>minProvisionedTPS</code>, track
54
- * your usage using Amazon CloudWatch metrics, and then increase the <code>minProvisionedTPS</code>
55
- * as necessary.</p>
54
+ * your usage using Amazon CloudWatch metrics, and then increase the <code>minProvisionedTPS</code>
55
+ * as necessary.</p>
56
56
  * <p>
57
57
  * <b>Status</b>
58
58
  * </p>
@@ -109,6 +109,7 @@ export interface CreateCampaignCommandOutput extends CreateCampaignResponse, __M
109
109
  * itemExplorationConfig: { // HyperParameters
110
110
  * "<keys>": "STRING_VALUE",
111
111
  * },
112
+ * enableMetadataWithRecommendations: true || false,
112
113
  * },
113
114
  * tags: [ // Tags
114
115
  * { // Tag
@@ -26,10 +26,10 @@ export interface CreateDatasetCommandOutput extends CreateDatasetResponse, __Met
26
26
  * <p>Creates an empty dataset and adds it to the specified dataset group.
27
27
  * Use <a href="https://docs.aws.amazon.com/personalize/latest/dg/API_CreateDatasetImportJob.html">CreateDatasetImportJob</a> to import your training data to a
28
28
  * dataset.</p>
29
- * <p>There are three types of datasets:</p>
29
+ * <p>There are 5 types of datasets:</p>
30
30
  * <ul>
31
31
  * <li>
32
- * <p>Interactions</p>
32
+ * <p>Item interactions</p>
33
33
  * </li>
34
34
  * <li>
35
35
  * <p>Items</p>
@@ -37,9 +37,15 @@ export interface CreateDatasetCommandOutput extends CreateDatasetResponse, __Met
37
37
  * <li>
38
38
  * <p>Users</p>
39
39
  * </li>
40
+ * <li>
41
+ * <p>Action interactions</p>
42
+ * </li>
43
+ * <li>
44
+ * <p>Actions</p>
45
+ * </li>
40
46
  * </ul>
41
47
  * <p>Each dataset type has an associated schema with required field types.
42
- * Only the <code>Interactions</code> dataset is required in order to train a
48
+ * Only the <code>Item interactions</code> dataset is required in order to train a
43
49
  * model (also referred to as creating a solution).</p>
44
50
  * <p>A dataset can be in one of the following states:</p>
45
51
  * <ul>
@@ -28,7 +28,7 @@ export interface CreateDatasetGroupCommandOutput extends CreateDatasetGroupRespo
28
28
  * for each type of dataset:</p>
29
29
  * <ul>
30
30
  * <li>
31
- * <p>Interactions</p>
31
+ * <p>Item interactions</p>
32
32
  * </li>
33
33
  * <li>
34
34
  * <p>Items</p>
@@ -36,6 +36,12 @@ export interface CreateDatasetGroupCommandOutput extends CreateDatasetGroupRespo
36
36
  * <li>
37
37
  * <p>Users</p>
38
38
  * </li>
39
+ * <li>
40
+ * <p>Actions</p>
41
+ * </li>
42
+ * <li>
43
+ * <p>Action interactions</p>
44
+ * </li>
39
45
  * </ul>
40
46
  * <p> A dataset group can be a Domain dataset group, where you specify a
41
47
  * domain and use pre-configured resources like recommenders, or a
@@ -30,6 +30,9 @@ export interface CreateDatasetImportJobCommandOutput extends CreateDatasetImport
30
30
  * processes it internally. For information on granting access to your Amazon S3
31
31
  * bucket, see <a href="https://docs.aws.amazon.com/personalize/latest/dg/granting-personalize-s3-access.html">Giving
32
32
  * Amazon Personalize Access to Amazon S3 Resources</a>. </p>
33
+ * <p>If you already created a recommender or deployed a custom solution version with a campaign, how new bulk records
34
+ * influence recommendations depends on the domain use case or recipe that you use. For more information, see <a href="https://docs.aws.amazon.com/personalize/latest/dg/how-new-data-influences-recommendations.html">How new data influences
35
+ * real-time recommendations</a>.</p>
33
36
  * <important>
34
37
  * <p>By default, a dataset import job replaces any existing data in the
35
38
  * dataset that you imported in bulk. To add new records without replacing
@@ -33,7 +33,7 @@ export interface CreateEventTrackerCommandOutput extends CreateEventTrackerRespo
33
33
  * </note>
34
34
  * <p>When you create an event tracker, the response includes a tracking ID, which you pass as a parameter when you use the
35
35
  * <a href="https://docs.aws.amazon.com/personalize/latest/dg/API_UBS_PutEvents.html">PutEvents</a> operation.
36
- * Amazon Personalize then appends the event data to the Interactions dataset of the dataset group you specify
36
+ * Amazon Personalize then appends the event data to the Item interactions dataset of the dataset group you specify
37
37
  * in your event tracker.
38
38
  * </p>
39
39
  * <p>The event tracker can be in one of the following states:</p>
@@ -123,6 +123,7 @@ export interface CreateRecommenderCommandOutput extends CreateRecommenderRespons
123
123
  * ],
124
124
  * },
125
125
  * },
126
+ * enableMetadataWithRecommendations: true || false,
126
127
  * },
127
128
  * tags: [ // Tags
128
129
  * { // Tag
@@ -23,8 +23,8 @@ export interface DeleteEventTrackerCommandOutput extends __MetadataBearer {
23
23
  }
24
24
  /**
25
25
  * @public
26
- * <p>Deletes the event tracker. Does not delete the event-interactions dataset from
27
- * the associated dataset group. For more
26
+ * <p>Deletes the event tracker. Does not delete the dataset from
27
+ * the dataset group. For more
28
28
  * information on event trackers, see <a href="https://docs.aws.amazon.com/personalize/latest/dg/API_CreateEventTracker.html">CreateEventTracker</a>.</p>
29
29
  * @example
30
30
  * Use a bare-bones client and the command you need to make an API call.
@@ -63,6 +63,12 @@ export interface DescribeBatchInferenceJobCommandOutput extends DescribeBatchInf
63
63
  * // },
64
64
  * // },
65
65
  * // roleArn: "STRING_VALUE",
66
+ * // batchInferenceJobMode: "BATCH_INFERENCE" || "THEME_GENERATION",
67
+ * // themeGenerationConfig: { // ThemeGenerationConfig
68
+ * // fieldsForThemeGeneration: { // FieldsForThemeGeneration
69
+ * // itemName: "STRING_VALUE", // required
70
+ * // },
71
+ * // },
66
72
  * // status: "STRING_VALUE",
67
73
  * // creationDateTime: new Date("TIMESTAMP"),
68
74
  * // lastUpdatedDateTime: new Date("TIMESTAMP"),
@@ -57,6 +57,7 @@ export interface DescribeCampaignCommandOutput extends DescribeCampaignResponse,
57
57
  * // itemExplorationConfig: { // HyperParameters
58
58
  * // "<keys>": "STRING_VALUE",
59
59
  * // },
60
+ * // enableMetadataWithRecommendations: true || false,
60
61
  * // },
61
62
  * // status: "STRING_VALUE",
62
63
  * // failureReason: "STRING_VALUE",
@@ -69,6 +70,7 @@ export interface DescribeCampaignCommandOutput extends DescribeCampaignResponse,
69
70
  * // itemExplorationConfig: {
70
71
  * // "<keys>": "STRING_VALUE",
71
72
  * // },
73
+ * // enableMetadataWithRecommendations: true || false,
72
74
  * // },
73
75
  * // status: "STRING_VALUE",
74
76
  * // failureReason: "STRING_VALUE",
@@ -53,6 +53,7 @@ export interface DescribeDatasetCommandOutput extends DescribeDatasetResponse, _
53
53
  * // creationDateTime: new Date("TIMESTAMP"),
54
54
  * // lastUpdatedDateTime: new Date("TIMESTAMP"),
55
55
  * // },
56
+ * // trackingId: "STRING_VALUE",
56
57
  * // },
57
58
  * // };
58
59
  *
@@ -70,6 +70,7 @@ export interface DescribeRecommenderCommandOutput extends DescribeRecommenderRes
70
70
  * // ],
71
71
  * // },
72
72
  * // },
73
+ * // enableMetadataWithRecommendations: true || false,
73
74
  * // },
74
75
  * // creationDateTime: new Date("TIMESTAMP"),
75
76
  * // lastUpdatedDateTime: new Date("TIMESTAMP"),
@@ -88,6 +89,7 @@ export interface DescribeRecommenderCommandOutput extends DescribeRecommenderRes
88
89
  * // ],
89
90
  * // },
90
91
  * // },
92
+ * // enableMetadataWithRecommendations: true || false,
91
93
  * // },
92
94
  * // creationDateTime: new Date("TIMESTAMP"),
93
95
  * // lastUpdatedDateTime: new Date("TIMESTAMP"),
@@ -48,6 +48,7 @@ export interface ListBatchInferenceJobsCommandOutput extends ListBatchInferenceJ
48
48
  * // lastUpdatedDateTime: new Date("TIMESTAMP"),
49
49
  * // failureReason: "STRING_VALUE",
50
50
  * // solutionVersionArn: "STRING_VALUE",
51
+ * // batchInferenceJobMode: "BATCH_INFERENCE" || "THEME_GENERATION",
51
52
  * // },
52
53
  * // ],
53
54
  * // nextToken: "STRING_VALUE",
@@ -59,6 +59,7 @@ export interface ListRecommendersCommandOutput extends ListRecommendersResponse,
59
59
  * // ],
60
60
  * // },
61
61
  * // },
62
+ * // enableMetadataWithRecommendations: true || false,
62
63
  * // },
63
64
  * // status: "STRING_VALUE",
64
65
  * // creationDateTime: new Date("TIMESTAMP"),
@@ -23,16 +23,19 @@ export interface UpdateCampaignCommandOutput extends UpdateCampaignResponse, __M
23
23
  }
24
24
  /**
25
25
  * @public
26
- * <p>Updates a campaign by either deploying a new solution or changing the value of the
27
- * campaign's <code>minProvisionedTPS</code> parameter.</p>
26
+ * <p>
27
+ * Updates a campaign to deploy a retrained solution version with an existing campaign, change your campaign's <code>minProvisionedTPS</code>,
28
+ * or modify your campaign's configuration, such as the exploration configuration.
29
+ * </p>
28
30
  * <p>To update a campaign, the campaign status must be ACTIVE or CREATE FAILED.
29
31
  * Check the campaign status using the <a href="https://docs.aws.amazon.com/personalize/latest/dg/API_DescribeCampaign.html">DescribeCampaign</a> operation.</p>
30
32
  * <note>
31
33
  * <p>You can still get recommendations from a campaign while an update is in progress.
32
- * The campaign will use the previous solution version and campaign configuration to generate recommendations until the latest campaign update status is <code>Active</code>.
33
- * </p>
34
+ * The campaign will use the previous solution version and campaign configuration to generate recommendations until the latest campaign update status is <code>Active</code>.
35
+ * </p>
34
36
  * </note>
35
- * <p>For more information on campaigns, see <a href="https://docs.aws.amazon.com/personalize/latest/dg/API_CreateCampaign.html">CreateCampaign</a>.</p>
37
+ * <p>For more information about updating a campaign, including code samples, see <a href="https://docs.aws.amazon.com/personalize/latest/dg/update-campaigns.html">Updating a campaign</a>.
38
+ * For more information about campaigns, see <a href="https://docs.aws.amazon.com/personalize/latest/dg/campaigns.html">Creating a campaign</a>.</p>
36
39
  * @example
37
40
  * Use a bare-bones client and the command you need to make an API call.
38
41
  * ```javascript
@@ -47,6 +50,7 @@ export interface UpdateCampaignCommandOutput extends UpdateCampaignResponse, __M
47
50
  * itemExplorationConfig: { // HyperParameters
48
51
  * "<keys>": "STRING_VALUE",
49
52
  * },
53
+ * enableMetadataWithRecommendations: true || false,
50
54
  * },
51
55
  * };
52
56
  * const command = new UpdateCampaignCommand(input);
@@ -51,6 +51,7 @@ export interface UpdateRecommenderCommandOutput extends UpdateRecommenderRespons
51
51
  * ],
52
52
  * },
53
53
  * },
54
+ * enableMetadataWithRecommendations: true || false,
54
55
  * },
55
56
  * };
56
57
  * const command = new UpdateRecommenderCommand(input);
@@ -189,6 +189,18 @@ export interface BatchInferenceJobConfig {
189
189
  */
190
190
  itemExplorationConfig?: Record<string, string>;
191
191
  }
192
+ /**
193
+ * @public
194
+ * @enum
195
+ */
196
+ export declare const BatchInferenceJobMode: {
197
+ readonly BATCH_INFERENCE: "BATCH_INFERENCE";
198
+ readonly THEME_GENERATION: "THEME_GENERATION";
199
+ };
200
+ /**
201
+ * @public
202
+ */
203
+ export type BatchInferenceJobMode = (typeof BatchInferenceJobMode)[keyof typeof BatchInferenceJobMode];
192
204
  /**
193
205
  * @public
194
206
  * <p>The configuration details of an Amazon S3 input or output bucket.</p>
@@ -247,6 +259,28 @@ export interface Tag {
247
259
  */
248
260
  tagValue: string | undefined;
249
261
  }
262
+ /**
263
+ * @public
264
+ * <p>A string to string map of the configuration details for theme generation.</p>
265
+ */
266
+ export interface FieldsForThemeGeneration {
267
+ /**
268
+ * @public
269
+ * <p>The name of the Items dataset column that stores the name of each item in the dataset.</p>
270
+ */
271
+ itemName: string | undefined;
272
+ }
273
+ /**
274
+ * @public
275
+ * <p>The configuration details for generating themes with a batch inference job.</p>
276
+ */
277
+ export interface ThemeGenerationConfig {
278
+ /**
279
+ * @public
280
+ * <p>Fields used to generate descriptive themes for a batch inference job.</p>
281
+ */
282
+ fieldsForThemeGeneration: FieldsForThemeGeneration | undefined;
283
+ }
250
284
  /**
251
285
  * @public
252
286
  */
@@ -301,6 +335,21 @@ export interface CreateBatchInferenceJobRequest {
301
335
  * <p>A list of <a href="https://docs.aws.amazon.com/personalize/latest/dg/tagging-resources.html">tags</a> to apply to the batch inference job.</p>
302
336
  */
303
337
  tags?: Tag[];
338
+ /**
339
+ * @public
340
+ * <p>The mode of the batch inference job. To generate descriptive themes for groups of similar items, set the
341
+ * job mode to <code>THEME_GENERATION</code>. If you don't want to generate themes, use the default <code>BATCH_INFERENCE</code>.</p>
342
+ * <p>
343
+ * When you get batch recommendations with themes, you will incur additional costs. For more information, see <a href="https://aws.amazon.com/personalize/pricing/">Amazon Personalize pricing</a>.
344
+ * </p>
345
+ */
346
+ batchInferenceJobMode?: BatchInferenceJobMode;
347
+ /**
348
+ * @public
349
+ * <p>For theme generation jobs, specify the name of the column in your Items
350
+ * dataset that contains each item's name.</p>
351
+ */
352
+ themeGenerationConfig?: ThemeGenerationConfig;
304
353
  }
305
354
  /**
306
355
  * @public
@@ -477,6 +526,15 @@ export interface CampaignConfig {
477
526
  * <a href="https://docs.aws.amazon.com/personalize/latest/dg/native-recipe-new-item-USER_PERSONALIZATION.html">User-Personalization</a> recipe.</p>
478
527
  */
479
528
  itemExplorationConfig?: Record<string, string>;
529
+ /**
530
+ * @public
531
+ * <p>Whether metadata with recommendations is enabled for the campaign.
532
+ * If enabled, you can specify the columns from your Items dataset in your request for recommendations. Amazon Personalize returns this data for each item in the recommendation response. </p>
533
+ * <p>
534
+ * If you enable metadata in recommendations, you will incur additional costs. For more information, see <a href="https://aws.amazon.com/personalize/pricing/">Amazon Personalize pricing</a>.
535
+ * </p>
536
+ */
537
+ enableMetadataWithRecommendations?: boolean;
480
538
  }
481
539
  /**
482
540
  * @public
@@ -555,6 +613,12 @@ export interface CreateDatasetRequest {
555
613
  * <li>
556
614
  * <p>Users</p>
557
615
  * </li>
616
+ * <li>
617
+ * <p>Actions</p>
618
+ * </li>
619
+ * <li>
620
+ * <p>Action_Interactions</p>
621
+ * </li>
558
622
  * </ul>
559
623
  */
560
624
  datasetType: string | undefined;
@@ -997,6 +1061,15 @@ export interface RecommenderConfig {
997
1061
  * </p>
998
1062
  */
999
1063
  trainingDataConfig?: TrainingDataConfig;
1064
+ /**
1065
+ * @public
1066
+ * <p>Whether metadata with recommendations is enabled for the recommender.
1067
+ * If enabled, you can specify the columns from your Items dataset in your request for recommendations. Amazon Personalize returns this data for each item in the recommendation response. </p>
1068
+ * <p>
1069
+ * If you enable metadata in recommendations, you will incur additional costs. For more information, see <a href="https://aws.amazon.com/personalize/pricing/">Amazon Personalize pricing</a>.
1070
+ * </p>
1071
+ */
1072
+ enableMetadataWithRecommendations?: boolean;
1000
1073
  }
1001
1074
  /**
1002
1075
  * @public
@@ -1295,7 +1368,7 @@ export interface SolutionConfig {
1295
1368
  hpoConfig?: HPOConfig;
1296
1369
  /**
1297
1370
  * @public
1298
- * <p>Lists the hyperparameter names and ranges.</p>
1371
+ * <p>Lists the algorithm hyperparameters and their values.</p>
1299
1372
  */
1300
1373
  algorithmHyperParameters?: Record<string, string>;
1301
1374
  /**
@@ -1432,19 +1505,25 @@ export interface CreateSolutionVersionRequest {
1432
1505
  solutionArn: string | undefined;
1433
1506
  /**
1434
1507
  * @public
1435
- * <p>The scope of training to be performed when creating the solution version. The
1436
- * <code>FULL</code> option trains the solution version based on the entirety of the input
1437
- * solution's training data, while the <code>UPDATE</code> option processes only the data that
1438
- * has changed in comparison to the input solution. Choose <code>UPDATE</code> when you want to
1439
- * incrementally update your solution version instead of creating an entirely new one.</p>
1440
- * <important>
1441
- * <p>The <code>UPDATE</code> option can only be used when you already have an active solution
1508
+ * <p>The scope of training to be performed when creating the solution version.
1509
+ * The default is <code>FULL</code>. This creates a completely new model based on the entirety
1510
+ * of the training data from the datasets in your dataset group.
1511
+ * </p>
1512
+ * <p>If you use
1513
+ * <a href="https://docs.aws.amazon.com/personalize/latest/dg/native-recipe-new-item-USER_PERSONALIZATION.html">User-Personalization</a>,
1514
+ * you can specify a training mode of <code>UPDATE</code>. This updates the model to consider new items for recommendations. It is not a full
1515
+ * retraining. You should still complete a full retraining weekly.
1516
+ * If you specify <code>UPDATE</code>, Amazon Personalize will stop automatic updates for the solution version. To resume updates, create a new solution with training mode set to <code>FULL</code>
1517
+ * and deploy it in a campaign.
1518
+ * For more information about automatic updates, see
1519
+ * <a href="https://docs.aws.amazon.com/personalize/latest/dg/use-case-recipe-features.html#maintaining-with-automatic-updates">Automatic updates</a>.
1520
+ * </p>
1521
+ * <p>The <code>UPDATE</code> option can only be used when you already have an active solution
1442
1522
  * version created from the input solution using the <code>FULL</code> option and the input
1443
1523
  * solution was trained with the
1444
1524
  * <a href="https://docs.aws.amazon.com/personalize/latest/dg/native-recipe-new-item-USER_PERSONALIZATION.html">User-Personalization</a>
1445
- * recipe or the
1525
+ * recipe or the legacy
1446
1526
  * <a href="https://docs.aws.amazon.com/personalize/latest/dg/native-recipe-hrnn-coldstart.html">HRNN-Coldstart</a> recipe.</p>
1447
- * </important>
1448
1527
  */
1449
1528
  trainingMode?: TrainingMode;
1450
1529
  /**
@@ -1641,6 +1720,16 @@ export interface BatchInferenceJob {
1641
1720
  * <p>The ARN of the Amazon Identity and Access Management (IAM) role that requested the batch inference job.</p>
1642
1721
  */
1643
1722
  roleArn?: string;
1723
+ /**
1724
+ * @public
1725
+ * <p>The job's mode.</p>
1726
+ */
1727
+ batchInferenceJobMode?: BatchInferenceJobMode;
1728
+ /**
1729
+ * @public
1730
+ * <p>The job's theme generation settings.</p>
1731
+ */
1732
+ themeGenerationConfig?: ThemeGenerationConfig;
1644
1733
  /**
1645
1734
  * @public
1646
1735
  * <p>The status of the batch inference job. The status is one of the following values:</p>
@@ -1997,6 +2086,12 @@ export interface Dataset {
1997
2086
  * <li>
1998
2087
  * <p>Users</p>
1999
2088
  * </li>
2089
+ * <li>
2090
+ * <p>Actions</p>
2091
+ * </li>
2092
+ * <li>
2093
+ * <p>Action_Interactions</p>
2094
+ * </li>
2000
2095
  * </ul>
2001
2096
  */
2002
2097
  datasetType?: string;
@@ -2035,6 +2130,12 @@ export interface Dataset {
2035
2130
  * <p>Describes the latest update to the dataset.</p>
2036
2131
  */
2037
2132
  latestDatasetUpdate?: DatasetUpdateSummary;
2133
+ /**
2134
+ * @public
2135
+ * <p>The ID of the event tracker for an Action interactions dataset.
2136
+ * You specify the tracker's ID in the <code>PutActionInteractions</code> API operation. Amazon Personalize uses it to direct new data to the Action interactions dataset in your dataset group.</p>
2137
+ */
2138
+ trackingId?: string;
2038
2139
  }
2039
2140
  /**
2040
2141
  * @public
@@ -2177,8 +2278,8 @@ export interface DescribeDatasetGroupRequest {
2177
2278
  }
2178
2279
  /**
2179
2280
  * @public
2180
- * <p>A dataset group is a collection of related datasets (Interactions,
2181
- * User, and Item). You create a dataset group by calling <a href="https://docs.aws.amazon.com/personalize/latest/dg/API_CreateDatasetGroup.html">CreateDatasetGroup</a>. You then create a dataset and add it to a
2281
+ * <p>A dataset group is a collection of related datasets (Item interactions,
2282
+ * Users, Items, Actions, Action interactions). You create a dataset group by calling <a href="https://docs.aws.amazon.com/personalize/latest/dg/API_CreateDatasetGroup.html">CreateDatasetGroup</a>. You then create a dataset and add it to a
2182
2283
  * dataset group by calling <a href="https://docs.aws.amazon.com/personalize/latest/dg/API_CreateDataset.html">CreateDataset</a>. The dataset group is used to create and train a
2183
2284
  * solution by calling <a href="https://docs.aws.amazon.com/personalize/latest/dg/API_CreateSolution.html">CreateSolution</a>. A dataset group can contain only one of each
2184
2285
  * type of dataset.</p>
@@ -2213,8 +2314,9 @@ export interface DatasetGroup {
2213
2314
  status?: string;
2214
2315
  /**
2215
2316
  * @public
2216
- * <p>The ARN of the IAM role that has permissions to create the dataset
2217
- * group.</p>
2317
+ * <p>The ARN of the Identity and Access Management (IAM) role that has permissions to access
2318
+ * the Key Management Service (KMS) key. Supplying an IAM role is only valid when also
2319
+ * specifying a KMS key.</p>
2218
2320
  */
2219
2321
  roleArn?: string;
2220
2322
  /**
@@ -3372,6 +3474,11 @@ export interface BatchInferenceJobSummary {
3372
3474
  * <p>The ARN of the solution version used by the batch inference job.</p>
3373
3475
  */
3374
3476
  solutionVersionArn?: string;
3477
+ /**
3478
+ * @public
3479
+ * <p>The job's mode.</p>
3480
+ */
3481
+ batchInferenceJobMode?: BatchInferenceJobMode;
3375
3482
  }
3376
3483
  /**
3377
3484
  * @public
@@ -3490,7 +3597,7 @@ export interface ListCampaignsRequest {
3490
3597
  /**
3491
3598
  * @public
3492
3599
  * <p>The Amazon Resource Name (ARN) of the solution to list the campaigns for. When
3493
- * a solution is not specified, all the campaigns associated with the account are listed.</p>
3600
+ * a solution is not specified, all the campaigns associated with the account are listed.</p>
3494
3601
  */
3495
3602
  solutionArn?: string;
3496
3603
  /**
@@ -3835,7 +3942,7 @@ export interface ListDatasetsRequest {
3835
3942
  /**
3836
3943
  * @public
3837
3944
  * <p>A token returned from the previous call to
3838
- * <code>ListDatasetImportJobs</code> for getting the next set of dataset
3945
+ * <code>ListDatasets</code> for getting the next set of dataset
3839
3946
  * import jobs (if they exist).</p>
3840
3947
  */
3841
3948
  nextToken?: string;