cdk-docker-image-deployment 0.0.87 → 0.0.88

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (60) hide show
  1. package/.jsii +3 -3
  2. package/lib/destination.js +1 -1
  3. package/lib/docker-image-deployment.js +1 -1
  4. package/lib/source.js +1 -1
  5. package/node_modules/aws-sdk/CHANGELOG.md +23 -1
  6. package/node_modules/aws-sdk/README.md +29 -1
  7. package/node_modules/aws-sdk/apis/backup-2018-11-15.min.json +441 -110
  8. package/node_modules/aws-sdk/apis/backup-2018-11-15.paginators.json +12 -0
  9. package/node_modules/aws-sdk/apis/drs-2020-02-26.min.json +124 -45
  10. package/node_modules/aws-sdk/apis/ecs-2014-11-13.min.json +331 -215
  11. package/node_modules/aws-sdk/apis/ecs-2014-11-13.paginators.json +6 -0
  12. package/node_modules/aws-sdk/apis/iot-2015-05-28.min.json +252 -214
  13. package/node_modules/aws-sdk/apis/iot-data-2015-05-28.min.json +26 -0
  14. package/node_modules/aws-sdk/apis/iotwireless-2020-11-22.min.json +686 -105
  15. package/node_modules/aws-sdk/apis/kendra-2019-02-03.min.json +43 -8
  16. package/node_modules/aws-sdk/apis/logs-2014-03-28.min.json +97 -16
  17. package/node_modules/aws-sdk/apis/metadata.json +3 -0
  18. package/node_modules/aws-sdk/apis/mgn-2020-02-26.min.json +931 -118
  19. package/node_modules/aws-sdk/apis/mgn-2020-02-26.paginators.json +24 -0
  20. package/node_modules/aws-sdk/apis/monitoring-2010-08-01.min.json +23 -14
  21. package/node_modules/aws-sdk/apis/monitoring-2010-08-01.paginators.json +4 -1
  22. package/node_modules/aws-sdk/apis/oam-2022-06-10.examples.json +5 -0
  23. package/node_modules/aws-sdk/apis/oam-2022-06-10.min.json +482 -0
  24. package/node_modules/aws-sdk/apis/oam-2022-06-10.paginators.json +22 -0
  25. package/node_modules/aws-sdk/apis/organizations-2016-11-28.min.json +62 -16
  26. package/node_modules/aws-sdk/apis/rds-2014-10-31.min.json +379 -236
  27. package/node_modules/aws-sdk/apis/rds-2014-10-31.paginators.json +6 -0
  28. package/node_modules/aws-sdk/apis/textract-2018-06-27.min.json +310 -62
  29. package/node_modules/aws-sdk/apis/transcribe-2017-10-26.min.json +98 -95
  30. package/node_modules/aws-sdk/clients/all.d.ts +1 -0
  31. package/node_modules/aws-sdk/clients/all.js +2 -1
  32. package/node_modules/aws-sdk/clients/backup.d.ts +405 -15
  33. package/node_modules/aws-sdk/clients/cloudwatch.d.ts +30 -16
  34. package/node_modules/aws-sdk/clients/cloudwatchlogs.d.ts +196 -65
  35. package/node_modules/aws-sdk/clients/drs.d.ts +104 -5
  36. package/node_modules/aws-sdk/clients/ecs.d.ts +159 -17
  37. package/node_modules/aws-sdk/clients/efs.d.ts +5 -5
  38. package/node_modules/aws-sdk/clients/iot.d.ts +76 -4
  39. package/node_modules/aws-sdk/clients/iotdata.d.ts +31 -1
  40. package/node_modules/aws-sdk/clients/iotwireless.d.ts +599 -14
  41. package/node_modules/aws-sdk/clients/kendra.d.ts +84 -39
  42. package/node_modules/aws-sdk/clients/mgn.d.ts +1046 -68
  43. package/node_modules/aws-sdk/clients/oam.d.ts +559 -0
  44. package/node_modules/aws-sdk/clients/oam.js +18 -0
  45. package/node_modules/aws-sdk/clients/organizations.d.ts +69 -8
  46. package/node_modules/aws-sdk/clients/rds.d.ts +199 -2
  47. package/node_modules/aws-sdk/clients/textract.d.ts +264 -0
  48. package/node_modules/aws-sdk/clients/transcribeservice.d.ts +210 -191
  49. package/node_modules/aws-sdk/dist/aws-sdk-core-react-native.js +29 -10
  50. package/node_modules/aws-sdk/dist/aws-sdk-react-native.js +91 -33
  51. package/node_modules/aws-sdk/dist/aws-sdk.js +1167 -707
  52. package/node_modules/aws-sdk/dist/aws-sdk.min.js +59 -59
  53. package/node_modules/aws-sdk/lib/config_service_placeholders.d.ts +2 -0
  54. package/node_modules/aws-sdk/lib/core.js +1 -1
  55. package/node_modules/aws-sdk/lib/event_listeners.js +22 -5
  56. package/node_modules/aws-sdk/lib/region_config.js +3 -1
  57. package/node_modules/aws-sdk/lib/region_config_data.json +11 -1
  58. package/node_modules/aws-sdk/lib/signers/bearer.js +1 -1
  59. package/node_modules/aws-sdk/package.json +1 -1
  60. package/package.json +6 -5
@@ -12,43 +12,43 @@ declare class TranscribeService extends Service {
12
12
  constructor(options?: TranscribeService.Types.ClientConfiguration)
13
13
  config: Config & TranscribeService.Types.ClientConfiguration;
14
14
  /**
15
- * Creates a new Call Analytics category. All categories are automatically applied to your Call Analytics jobs. Note that in order to apply your categories to your jobs, you must create them before submitting your job request, as categories cannot be applied retroactively. Call Analytics categories are composed of rules. For each category, you must create between 1 and 20 rules. Rules can include these parameters: , , , and . To update an existing category, see . To learn more about: Call Analytics categories, see Creating categories Using rules, see Rule criteria and refer to the data type Call Analytics, see Analyzing call center audio with Call Analytics
15
+ * Creates a new Call Analytics category. All categories are automatically applied to your Call Analytics transcriptions. Note that in order to apply categories to your transcriptions, you must create them before submitting your transcription request, as categories cannot be applied retroactively. When creating a new category, you can use the InputType parameter to label the category as a batch category (POST_CALL) or a streaming category (REAL_TIME). Batch categories can only be applied to batch transcriptions and streaming categories can only be applied to streaming transcriptions. If you do not include InputType, your category is created as a batch category by default. Call Analytics categories are composed of rules. For each category, you must create between 1 and 20 rules. Rules can include these parameters: , , , and . To update an existing category, see . To learn more about Call Analytics categories, see Creating categories for batch transcriptions and Creating categories for streaming transcriptions.
16
16
  */
17
17
  createCallAnalyticsCategory(params: TranscribeService.Types.CreateCallAnalyticsCategoryRequest, callback?: (err: AWSError, data: TranscribeService.Types.CreateCallAnalyticsCategoryResponse) => void): Request<TranscribeService.Types.CreateCallAnalyticsCategoryResponse, AWSError>;
18
18
  /**
19
- * Creates a new Call Analytics category. All categories are automatically applied to your Call Analytics jobs. Note that in order to apply your categories to your jobs, you must create them before submitting your job request, as categories cannot be applied retroactively. Call Analytics categories are composed of rules. For each category, you must create between 1 and 20 rules. Rules can include these parameters: , , , and . To update an existing category, see . To learn more about: Call Analytics categories, see Creating categories Using rules, see Rule criteria and refer to the data type Call Analytics, see Analyzing call center audio with Call Analytics
19
+ * Creates a new Call Analytics category. All categories are automatically applied to your Call Analytics transcriptions. Note that in order to apply categories to your transcriptions, you must create them before submitting your transcription request, as categories cannot be applied retroactively. When creating a new category, you can use the InputType parameter to label the category as a batch category (POST_CALL) or a streaming category (REAL_TIME). Batch categories can only be applied to batch transcriptions and streaming categories can only be applied to streaming transcriptions. If you do not include InputType, your category is created as a batch category by default. Call Analytics categories are composed of rules. For each category, you must create between 1 and 20 rules. Rules can include these parameters: , , , and . To update an existing category, see . To learn more about Call Analytics categories, see Creating categories for batch transcriptions and Creating categories for streaming transcriptions.
20
20
  */
21
21
  createCallAnalyticsCategory(callback?: (err: AWSError, data: TranscribeService.Types.CreateCallAnalyticsCategoryResponse) => void): Request<TranscribeService.Types.CreateCallAnalyticsCategoryResponse, AWSError>;
22
22
  /**
23
- * Creates a new custom language model. When creating a new language model, you must specify: If you want a Wideband (audio sample rates over 16,000 Hz) or Narrowband (audio sample rates under 16,000 Hz) base model The location of your training and tuning files (this must be an Amazon S3 URI) The language of your model A unique name for your model For more information, see Custom language models.
23
+ * Creates a new custom language model. When creating a new custom language model, you must specify: If you want a Wideband (audio sample rates over 16,000 Hz) or Narrowband (audio sample rates under 16,000 Hz) base model The location of your training and tuning files (this must be an Amazon S3 URI) The language of your model A unique name for your model
24
24
  */
25
25
  createLanguageModel(params: TranscribeService.Types.CreateLanguageModelRequest, callback?: (err: AWSError, data: TranscribeService.Types.CreateLanguageModelResponse) => void): Request<TranscribeService.Types.CreateLanguageModelResponse, AWSError>;
26
26
  /**
27
- * Creates a new custom language model. When creating a new language model, you must specify: If you want a Wideband (audio sample rates over 16,000 Hz) or Narrowband (audio sample rates under 16,000 Hz) base model The location of your training and tuning files (this must be an Amazon S3 URI) The language of your model A unique name for your model For more information, see Custom language models.
27
+ * Creates a new custom language model. When creating a new custom language model, you must specify: If you want a Wideband (audio sample rates over 16,000 Hz) or Narrowband (audio sample rates under 16,000 Hz) base model The location of your training and tuning files (this must be an Amazon S3 URI) The language of your model A unique name for your model
28
28
  */
29
29
  createLanguageModel(callback?: (err: AWSError, data: TranscribeService.Types.CreateLanguageModelResponse) => void): Request<TranscribeService.Types.CreateLanguageModelResponse, AWSError>;
30
30
  /**
31
- * Creates a new custom medical vocabulary. Prior to creating a new medical vocabulary, you must first upload a text file that contains your new entries, phrases, and terms into an Amazon S3 bucket. Note that this differs from , where you can include a list of terms within your request using the Phrases flag; CreateMedicalVocabulary does not support the Phrases flag. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your vocabulary request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language. For more information, see Creating a custom vocabulary.
31
+ * Creates a new custom medical vocabulary. Before creating a new custom medical vocabulary, you must first upload a text file that contains your new entries, phrases, and terms into an Amazon S3 bucket. Note that this differs from , where you can include a list of terms within your request using the Phrases flag; CreateMedicalVocabulary does not support the Phrases flag. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language. For more information, see Custom vocabularies.
32
32
  */
33
33
  createMedicalVocabulary(params: TranscribeService.Types.CreateMedicalVocabularyRequest, callback?: (err: AWSError, data: TranscribeService.Types.CreateMedicalVocabularyResponse) => void): Request<TranscribeService.Types.CreateMedicalVocabularyResponse, AWSError>;
34
34
  /**
35
- * Creates a new custom medical vocabulary. Prior to creating a new medical vocabulary, you must first upload a text file that contains your new entries, phrases, and terms into an Amazon S3 bucket. Note that this differs from , where you can include a list of terms within your request using the Phrases flag; CreateMedicalVocabulary does not support the Phrases flag. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your vocabulary request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language. For more information, see Creating a custom vocabulary.
35
+ * Creates a new custom medical vocabulary. Before creating a new custom medical vocabulary, you must first upload a text file that contains your new entries, phrases, and terms into an Amazon S3 bucket. Note that this differs from , where you can include a list of terms within your request using the Phrases flag; CreateMedicalVocabulary does not support the Phrases flag. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language. For more information, see Custom vocabularies.
36
36
  */
37
37
  createMedicalVocabulary(callback?: (err: AWSError, data: TranscribeService.Types.CreateMedicalVocabularyResponse) => void): Request<TranscribeService.Types.CreateMedicalVocabularyResponse, AWSError>;
38
38
  /**
39
- * Creates a new custom vocabulary. When creating a new vocabulary, you can either upload a text file that contains your new entries, phrases, and terms into an Amazon S3 bucket and include the URI in your request, or you can include a list of terms directly in your request using the Phrases flag. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your vocabulary request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language. For more information, see Creating a custom vocabulary.
39
+ * Creates a new custom vocabulary. When creating a new custom vocabulary, you can either upload a text file that contains your new entries, phrases, and terms into an Amazon S3 bucket and include the URI in your request. Or you can include a list of terms directly in your request using the Phrases flag. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language. For more information, see Custom vocabularies.
40
40
  */
41
41
  createVocabulary(params: TranscribeService.Types.CreateVocabularyRequest, callback?: (err: AWSError, data: TranscribeService.Types.CreateVocabularyResponse) => void): Request<TranscribeService.Types.CreateVocabularyResponse, AWSError>;
42
42
  /**
43
- * Creates a new custom vocabulary. When creating a new vocabulary, you can either upload a text file that contains your new entries, phrases, and terms into an Amazon S3 bucket and include the URI in your request, or you can include a list of terms directly in your request using the Phrases flag. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your vocabulary request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language. For more information, see Creating a custom vocabulary.
43
+ * Creates a new custom vocabulary. When creating a new custom vocabulary, you can either upload a text file that contains your new entries, phrases, and terms into an Amazon S3 bucket and include the URI in your request. Or you can include a list of terms directly in your request using the Phrases flag. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language. For more information, see Custom vocabularies.
44
44
  */
45
45
  createVocabulary(callback?: (err: AWSError, data: TranscribeService.Types.CreateVocabularyResponse) => void): Request<TranscribeService.Types.CreateVocabularyResponse, AWSError>;
46
46
  /**
47
- * Creates a new custom vocabulary filter. You can use vocabulary filters to mask, delete, or flag specific words from your transcript. Vocabulary filters are commonly used to mask profanity in transcripts. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language. For more information, see Using vocabulary filtering with unwanted words.
47
+ * Creates a new custom vocabulary filter. You can use custom vocabulary filters to mask, delete, or flag specific words from your transcript. Custom vocabulary filters are commonly used to mask profanity in transcripts. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language. For more information, see Vocabulary filtering.
48
48
  */
49
49
  createVocabularyFilter(params: TranscribeService.Types.CreateVocabularyFilterRequest, callback?: (err: AWSError, data: TranscribeService.Types.CreateVocabularyFilterResponse) => void): Request<TranscribeService.Types.CreateVocabularyFilterResponse, AWSError>;
50
50
  /**
51
- * Creates a new custom vocabulary filter. You can use vocabulary filters to mask, delete, or flag specific words from your transcript. Vocabulary filters are commonly used to mask profanity in transcripts. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language. For more information, see Using vocabulary filtering with unwanted words.
51
+ * Creates a new custom vocabulary filter. You can use custom vocabulary filters to mask, delete, or flag specific words from your transcript. Custom vocabulary filters are commonly used to mask profanity in transcripts. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language. For more information, see Vocabulary filtering.
52
52
  */
53
53
  createVocabularyFilter(callback?: (err: AWSError, data: TranscribeService.Types.CreateVocabularyFilterResponse) => void): Request<TranscribeService.Types.CreateVocabularyFilterResponse, AWSError>;
54
54
  /**
@@ -68,11 +68,11 @@ declare class TranscribeService extends Service {
68
68
  */
69
69
  deleteCallAnalyticsJob(callback?: (err: AWSError, data: TranscribeService.Types.DeleteCallAnalyticsJobResponse) => void): Request<TranscribeService.Types.DeleteCallAnalyticsJobResponse, AWSError>;
70
70
  /**
71
- * Deletes a custom language model. To use this operation, specify the name of the language model you want to delete using ModelName. Language model names are case sensitive.
71
+ * Deletes a custom language model. To use this operation, specify the name of the language model you want to delete using ModelName. custom language model names are case sensitive.
72
72
  */
73
73
  deleteLanguageModel(params: TranscribeService.Types.DeleteLanguageModelRequest, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
74
74
  /**
75
- * Deletes a custom language model. To use this operation, specify the name of the language model you want to delete using ModelName. Language model names are case sensitive.
75
+ * Deletes a custom language model. To use this operation, specify the name of the language model you want to delete using ModelName. custom language model names are case sensitive.
76
76
  */
77
77
  deleteLanguageModel(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
78
78
  /**
@@ -84,11 +84,11 @@ declare class TranscribeService extends Service {
84
84
  */
85
85
  deleteMedicalTranscriptionJob(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
86
86
  /**
87
- * Deletes a custom medical vocabulary. To use this operation, specify the name of the vocabulary you want to delete using VocabularyName. Vocabulary names are case sensitive.
87
+ * Deletes a custom medical vocabulary. To use this operation, specify the name of the custom vocabulary you want to delete using VocabularyName. Custom vocabulary names are case sensitive.
88
88
  */
89
89
  deleteMedicalVocabulary(params: TranscribeService.Types.DeleteMedicalVocabularyRequest, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
90
90
  /**
91
- * Deletes a custom medical vocabulary. To use this operation, specify the name of the vocabulary you want to delete using VocabularyName. Vocabulary names are case sensitive.
91
+ * Deletes a custom medical vocabulary. To use this operation, specify the name of the custom vocabulary you want to delete using VocabularyName. Custom vocabulary names are case sensitive.
92
92
  */
93
93
  deleteMedicalVocabulary(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
94
94
  /**
@@ -100,27 +100,27 @@ declare class TranscribeService extends Service {
100
100
  */
101
101
  deleteTranscriptionJob(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
102
102
  /**
103
- * Deletes a custom vocabulary. To use this operation, specify the name of the vocabulary you want to delete using VocabularyName. Vocabulary names are case sensitive.
103
+ * Deletes a custom vocabulary. To use this operation, specify the name of the custom vocabulary you want to delete using VocabularyName. Custom vocabulary names are case sensitive.
104
104
  */
105
105
  deleteVocabulary(params: TranscribeService.Types.DeleteVocabularyRequest, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
106
106
  /**
107
- * Deletes a custom vocabulary. To use this operation, specify the name of the vocabulary you want to delete using VocabularyName. Vocabulary names are case sensitive.
107
+ * Deletes a custom vocabulary. To use this operation, specify the name of the custom vocabulary you want to delete using VocabularyName. Custom vocabulary names are case sensitive.
108
108
  */
109
109
  deleteVocabulary(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
110
110
  /**
111
- * Deletes a vocabulary filter. To use this operation, specify the name of the vocabulary filter you want to delete using VocabularyFilterName. Vocabulary filter names are case sensitive.
111
+ * Deletes a custom vocabulary filter. To use this operation, specify the name of the custom vocabulary filter you want to delete using VocabularyFilterName. Custom vocabulary filter names are case sensitive.
112
112
  */
113
113
  deleteVocabularyFilter(params: TranscribeService.Types.DeleteVocabularyFilterRequest, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
114
114
  /**
115
- * Deletes a vocabulary filter. To use this operation, specify the name of the vocabulary filter you want to delete using VocabularyFilterName. Vocabulary filter names are case sensitive.
115
+ * Deletes a custom vocabulary filter. To use this operation, specify the name of the custom vocabulary filter you want to delete using VocabularyFilterName. Custom vocabulary filter names are case sensitive.
116
116
  */
117
117
  deleteVocabularyFilter(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
118
118
  /**
119
- * Provides information about the specified custom language model. This operation also shows if the base language model you used to create your custom language model has been updated. If Amazon Transcribe has updated the base model, you can create a new custom language model using the updated base model. If you tried to create a new custom language model and the request wasn't successful, you can use DescribeLanguageModel to help identify the reason for this failure. To get a list of your custom language models, use the operation.
119
+ * Provides information about the specified custom language model. This operation also shows if the base language model that you used to create your custom language model has been updated. If Amazon Transcribe has updated the base model, you can create a new custom language model using the updated base model. If you tried to create a new custom language model and the request wasn't successful, you can use DescribeLanguageModel to help identify the reason for this failure.
120
120
  */
121
121
  describeLanguageModel(params: TranscribeService.Types.DescribeLanguageModelRequest, callback?: (err: AWSError, data: TranscribeService.Types.DescribeLanguageModelResponse) => void): Request<TranscribeService.Types.DescribeLanguageModelResponse, AWSError>;
122
122
  /**
123
- * Provides information about the specified custom language model. This operation also shows if the base language model you used to create your custom language model has been updated. If Amazon Transcribe has updated the base model, you can create a new custom language model using the updated base model. If you tried to create a new custom language model and the request wasn't successful, you can use DescribeLanguageModel to help identify the reason for this failure. To get a list of your custom language models, use the operation.
123
+ * Provides information about the specified custom language model. This operation also shows if the base language model that you used to create your custom language model has been updated. If Amazon Transcribe has updated the base model, you can create a new custom language model using the updated base model. If you tried to create a new custom language model and the request wasn't successful, you can use DescribeLanguageModel to help identify the reason for this failure.
124
124
  */
125
125
  describeLanguageModel(callback?: (err: AWSError, data: TranscribeService.Types.DescribeLanguageModelResponse) => void): Request<TranscribeService.Types.DescribeLanguageModelResponse, AWSError>;
126
126
  /**
@@ -140,43 +140,43 @@ declare class TranscribeService extends Service {
140
140
  */
141
141
  getCallAnalyticsJob(callback?: (err: AWSError, data: TranscribeService.Types.GetCallAnalyticsJobResponse) => void): Request<TranscribeService.Types.GetCallAnalyticsJobResponse, AWSError>;
142
142
  /**
143
- * Provides information about the specified medical transcription job. To view the status of the specified medical transcription job, check the TranscriptionJobStatus field. If the status is COMPLETED, the job is finished and you can find the results at the location specified in TranscriptFileUri. If the status is FAILED, FailureReason provides details on why your transcription job failed. To get a list of your medical transcription jobs, use the operation.
143
+ * Provides information about the specified medical transcription job. To view the status of the specified medical transcription job, check the TranscriptionJobStatus field. If the status is COMPLETED, the job is finished. You can find the results at the location specified in TranscriptFileUri. If the status is FAILED, FailureReason provides details on why your transcription job failed. To get a list of your medical transcription jobs, use the operation.
144
144
  */
145
145
  getMedicalTranscriptionJob(params: TranscribeService.Types.GetMedicalTranscriptionJobRequest, callback?: (err: AWSError, data: TranscribeService.Types.GetMedicalTranscriptionJobResponse) => void): Request<TranscribeService.Types.GetMedicalTranscriptionJobResponse, AWSError>;
146
146
  /**
147
- * Provides information about the specified medical transcription job. To view the status of the specified medical transcription job, check the TranscriptionJobStatus field. If the status is COMPLETED, the job is finished and you can find the results at the location specified in TranscriptFileUri. If the status is FAILED, FailureReason provides details on why your transcription job failed. To get a list of your medical transcription jobs, use the operation.
147
+ * Provides information about the specified medical transcription job. To view the status of the specified medical transcription job, check the TranscriptionJobStatus field. If the status is COMPLETED, the job is finished. You can find the results at the location specified in TranscriptFileUri. If the status is FAILED, FailureReason provides details on why your transcription job failed. To get a list of your medical transcription jobs, use the operation.
148
148
  */
149
149
  getMedicalTranscriptionJob(callback?: (err: AWSError, data: TranscribeService.Types.GetMedicalTranscriptionJobResponse) => void): Request<TranscribeService.Types.GetMedicalTranscriptionJobResponse, AWSError>;
150
150
  /**
151
- * Provides information about the specified custom medical vocabulary. To view the status of the specified medical vocabulary, check the VocabularyState field. If the status is READY, your vocabulary is available to use. If the status is FAILED, FailureReason provides details on why your vocabulary failed. To get a list of your custom medical vocabularies, use the operation.
151
+ * Provides information about the specified custom medical vocabulary. To view the status of the specified custom medical vocabulary, check the VocabularyState field. If the status is READY, your custom vocabulary is available to use. If the status is FAILED, FailureReason provides details on why your vocabulary failed. To get a list of your custom medical vocabularies, use the operation.
152
152
  */
153
153
  getMedicalVocabulary(params: TranscribeService.Types.GetMedicalVocabularyRequest, callback?: (err: AWSError, data: TranscribeService.Types.GetMedicalVocabularyResponse) => void): Request<TranscribeService.Types.GetMedicalVocabularyResponse, AWSError>;
154
154
  /**
155
- * Provides information about the specified custom medical vocabulary. To view the status of the specified medical vocabulary, check the VocabularyState field. If the status is READY, your vocabulary is available to use. If the status is FAILED, FailureReason provides details on why your vocabulary failed. To get a list of your custom medical vocabularies, use the operation.
155
+ * Provides information about the specified custom medical vocabulary. To view the status of the specified custom medical vocabulary, check the VocabularyState field. If the status is READY, your custom vocabulary is available to use. If the status is FAILED, FailureReason provides details on why your vocabulary failed. To get a list of your custom medical vocabularies, use the operation.
156
156
  */
157
157
  getMedicalVocabulary(callback?: (err: AWSError, data: TranscribeService.Types.GetMedicalVocabularyResponse) => void): Request<TranscribeService.Types.GetMedicalVocabularyResponse, AWSError>;
158
158
  /**
159
- * Provides information about the specified transcription job. To view the status of the specified transcription job, check the TranscriptionJobStatus field. If the status is COMPLETED, the job is finished and you can find the results at the location specified in TranscriptFileUri. If the status is FAILED, FailureReason provides details on why your transcription job failed. If you enabled content redaction, the redacted transcript can be found at the location specified in RedactedTranscriptFileUri. To get a list of your transcription jobs, use the operation.
159
+ * Provides information about the specified transcription job. To view the status of the specified transcription job, check the TranscriptionJobStatus field. If the status is COMPLETED, the job is finished. You can find the results at the location specified in TranscriptFileUri. If the status is FAILED, FailureReason provides details on why your transcription job failed. If you enabled content redaction, the redacted transcript can be found at the location specified in RedactedTranscriptFileUri. To get a list of your transcription jobs, use the operation.
160
160
  */
161
161
  getTranscriptionJob(params: TranscribeService.Types.GetTranscriptionJobRequest, callback?: (err: AWSError, data: TranscribeService.Types.GetTranscriptionJobResponse) => void): Request<TranscribeService.Types.GetTranscriptionJobResponse, AWSError>;
162
162
  /**
163
- * Provides information about the specified transcription job. To view the status of the specified transcription job, check the TranscriptionJobStatus field. If the status is COMPLETED, the job is finished and you can find the results at the location specified in TranscriptFileUri. If the status is FAILED, FailureReason provides details on why your transcription job failed. If you enabled content redaction, the redacted transcript can be found at the location specified in RedactedTranscriptFileUri. To get a list of your transcription jobs, use the operation.
163
+ * Provides information about the specified transcription job. To view the status of the specified transcription job, check the TranscriptionJobStatus field. If the status is COMPLETED, the job is finished. You can find the results at the location specified in TranscriptFileUri. If the status is FAILED, FailureReason provides details on why your transcription job failed. If you enabled content redaction, the redacted transcript can be found at the location specified in RedactedTranscriptFileUri. To get a list of your transcription jobs, use the operation.
164
164
  */
165
165
  getTranscriptionJob(callback?: (err: AWSError, data: TranscribeService.Types.GetTranscriptionJobResponse) => void): Request<TranscribeService.Types.GetTranscriptionJobResponse, AWSError>;
166
166
  /**
167
- * Provides information about the specified custom vocabulary. To view the status of the specified vocabulary, check the VocabularyState field. If the status is READY, your vocabulary is available to use. If the status is FAILED, FailureReason provides details on why your vocabulary failed. To get a list of your custom vocabularies, use the operation.
167
+ * Provides information about the specified custom vocabulary. To view the status of the specified custom vocabulary, check the VocabularyState field. If the status is READY, your custom vocabulary is available to use. If the status is FAILED, FailureReason provides details on why your custom vocabulary failed. To get a list of your custom vocabularies, use the operation.
168
168
  */
169
169
  getVocabulary(params: TranscribeService.Types.GetVocabularyRequest, callback?: (err: AWSError, data: TranscribeService.Types.GetVocabularyResponse) => void): Request<TranscribeService.Types.GetVocabularyResponse, AWSError>;
170
170
  /**
171
- * Provides information about the specified custom vocabulary. To view the status of the specified vocabulary, check the VocabularyState field. If the status is READY, your vocabulary is available to use. If the status is FAILED, FailureReason provides details on why your vocabulary failed. To get a list of your custom vocabularies, use the operation.
171
+ * Provides information about the specified custom vocabulary. To view the status of the specified custom vocabulary, check the VocabularyState field. If the status is READY, your custom vocabulary is available to use. If the status is FAILED, FailureReason provides details on why your custom vocabulary failed. To get a list of your custom vocabularies, use the operation.
172
172
  */
173
173
  getVocabulary(callback?: (err: AWSError, data: TranscribeService.Types.GetVocabularyResponse) => void): Request<TranscribeService.Types.GetVocabularyResponse, AWSError>;
174
174
  /**
175
- * Provides information about the specified custom vocabulary filter. To view the status of the specified vocabulary filter, check the VocabularyState field. If the status is READY, your vocabulary is available to use. If the status is FAILED, FailureReason provides details on why your vocabulary filter failed. To get a list of your custom vocabulary filters, use the operation.
175
+ * Provides information about the specified custom vocabulary filter. To get a list of your custom vocabulary filters, use the operation.
176
176
  */
177
177
  getVocabularyFilter(params: TranscribeService.Types.GetVocabularyFilterRequest, callback?: (err: AWSError, data: TranscribeService.Types.GetVocabularyFilterResponse) => void): Request<TranscribeService.Types.GetVocabularyFilterResponse, AWSError>;
178
178
  /**
179
- * Provides information about the specified custom vocabulary filter. To view the status of the specified vocabulary filter, check the VocabularyState field. If the status is READY, your vocabulary is available to use. If the status is FAILED, FailureReason provides details on why your vocabulary filter failed. To get a list of your custom vocabulary filters, use the operation.
179
+ * Provides information about the specified custom vocabulary filter. To get a list of your custom vocabulary filters, use the operation.
180
180
  */
181
181
  getVocabularyFilter(callback?: (err: AWSError, data: TranscribeService.Types.GetVocabularyFilterResponse) => void): Request<TranscribeService.Types.GetVocabularyFilterResponse, AWSError>;
182
182
  /**
@@ -196,11 +196,11 @@ declare class TranscribeService extends Service {
196
196
  */
197
197
  listCallAnalyticsJobs(callback?: (err: AWSError, data: TranscribeService.Types.ListCallAnalyticsJobsResponse) => void): Request<TranscribeService.Types.ListCallAnalyticsJobsResponse, AWSError>;
198
198
  /**
199
- * Provides a list of custom language models that match the specified criteria. If no criteria are specified, all language models are returned. To get detailed information about a specific custom language model, use the operation.
199
+ * Provides a list of custom language models that match the specified criteria. If no criteria are specified, all custom language models are returned. To get detailed information about a specific custom language model, use the operation.
200
200
  */
201
201
  listLanguageModels(params: TranscribeService.Types.ListLanguageModelsRequest, callback?: (err: AWSError, data: TranscribeService.Types.ListLanguageModelsResponse) => void): Request<TranscribeService.Types.ListLanguageModelsResponse, AWSError>;
202
202
  /**
203
- * Provides a list of custom language models that match the specified criteria. If no criteria are specified, all language models are returned. To get detailed information about a specific custom language model, use the operation.
203
+ * Provides a list of custom language models that match the specified criteria. If no criteria are specified, all custom language models are returned. To get detailed information about a specific custom language model, use the operation.
204
204
  */
205
205
  listLanguageModels(callback?: (err: AWSError, data: TranscribeService.Types.ListLanguageModelsResponse) => void): Request<TranscribeService.Types.ListLanguageModelsResponse, AWSError>;
206
206
  /**
@@ -252,19 +252,19 @@ declare class TranscribeService extends Service {
252
252
  */
253
253
  listVocabularyFilters(callback?: (err: AWSError, data: TranscribeService.Types.ListVocabularyFiltersResponse) => void): Request<TranscribeService.Types.ListVocabularyFiltersResponse, AWSError>;
254
254
  /**
255
- * Transcribes the audio from a customer service call and applies any additional Request Parameters you choose to include in your request. In addition to many of the standard transcription features, Call Analytics provides you with call characteristics, call summarization, speaker sentiment, and optional redaction of your text transcript and your audio file. You can also apply custom categories to flag specified conditions. To learn more about these features and insights, refer to Analyzing call center audio with Call Analytics. If you want to apply categories to your Call Analytics job, you must create them before submitting your job request. Categories cannot be retroactively applied to a job. To create a new category, use the operation. To learn more about Call Analytics categories, see Creating categories. To make a StartCallAnalyticsJob request, you must first upload your media file into an Amazon S3 bucket; you can then specify the Amazon S3 location of the file using the Media parameter. You must include the following parameters in your StartCallAnalyticsJob request: region: The Amazon Web Services Region where you are making your request. For a list of Amazon Web Services Regions supported with Amazon Transcribe, refer to Amazon Transcribe endpoints and quotas. CallAnalyticsJobName: A custom name you create for your transcription job that is unique within your Amazon Web Services account. DataAccessRoleArn: The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. Media (MediaFileUri or RedactedMediaFileUri): The Amazon S3 location of your media file. With Call Analytics, you can redact the audio contained in your media file by including RedactedMediaFileUri, instead of MediaFileUri, to specify the location of your input audio. If you choose to redact your audio, you can find your redacted media at the location specified in the RedactedMediaFileUri field of your response.
255
+ * Transcribes the audio from a customer service call and applies any additional Request Parameters you choose to include in your request. In addition to many standard transcription features, Call Analytics provides you with call characteristics, call summarization, speaker sentiment, and optional redaction of your text transcript and your audio file. You can also apply custom categories to flag specified conditions. To learn more about these features and insights, refer to Analyzing call center audio with Call Analytics. If you want to apply categories to your Call Analytics job, you must create them before submitting your job request. Categories cannot be retroactively applied to a job. To create a new category, use the operation. To learn more about Call Analytics categories, see Creating categories for batch transcriptions and Creating categories for streaming transcriptions. To make a StartCallAnalyticsJob request, you must first upload your media file into an Amazon S3 bucket; you can then specify the Amazon S3 location of the file using the Media parameter. You must include the following parameters in your StartCallAnalyticsJob request: region: The Amazon Web Services Region where you are making your request. For a list of Amazon Web Services Regions supported with Amazon Transcribe, refer to Amazon Transcribe endpoints and quotas. CallAnalyticsJobName: A custom name that you create for your transcription job that's unique within your Amazon Web Services account. DataAccessRoleArn: The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. Media (MediaFileUri or RedactedMediaFileUri): The Amazon S3 location of your media file. With Call Analytics, you can redact the audio contained in your media file by including RedactedMediaFileUri, instead of MediaFileUri, to specify the location of your input audio. If you choose to redact your audio, you can find your redacted media at the location specified in the RedactedMediaFileUri field of your response.
256
256
  */
257
257
  startCallAnalyticsJob(params: TranscribeService.Types.StartCallAnalyticsJobRequest, callback?: (err: AWSError, data: TranscribeService.Types.StartCallAnalyticsJobResponse) => void): Request<TranscribeService.Types.StartCallAnalyticsJobResponse, AWSError>;
258
258
  /**
259
- * Transcribes the audio from a customer service call and applies any additional Request Parameters you choose to include in your request. In addition to many of the standard transcription features, Call Analytics provides you with call characteristics, call summarization, speaker sentiment, and optional redaction of your text transcript and your audio file. You can also apply custom categories to flag specified conditions. To learn more about these features and insights, refer to Analyzing call center audio with Call Analytics. If you want to apply categories to your Call Analytics job, you must create them before submitting your job request. Categories cannot be retroactively applied to a job. To create a new category, use the operation. To learn more about Call Analytics categories, see Creating categories. To make a StartCallAnalyticsJob request, you must first upload your media file into an Amazon S3 bucket; you can then specify the Amazon S3 location of the file using the Media parameter. You must include the following parameters in your StartCallAnalyticsJob request: region: The Amazon Web Services Region where you are making your request. For a list of Amazon Web Services Regions supported with Amazon Transcribe, refer to Amazon Transcribe endpoints and quotas. CallAnalyticsJobName: A custom name you create for your transcription job that is unique within your Amazon Web Services account. DataAccessRoleArn: The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. Media (MediaFileUri or RedactedMediaFileUri): The Amazon S3 location of your media file. With Call Analytics, you can redact the audio contained in your media file by including RedactedMediaFileUri, instead of MediaFileUri, to specify the location of your input audio. If you choose to redact your audio, you can find your redacted media at the location specified in the RedactedMediaFileUri field of your response.
259
+ * Transcribes the audio from a customer service call and applies any additional Request Parameters you choose to include in your request. In addition to many standard transcription features, Call Analytics provides you with call characteristics, call summarization, speaker sentiment, and optional redaction of your text transcript and your audio file. You can also apply custom categories to flag specified conditions. To learn more about these features and insights, refer to Analyzing call center audio with Call Analytics. If you want to apply categories to your Call Analytics job, you must create them before submitting your job request. Categories cannot be retroactively applied to a job. To create a new category, use the operation. To learn more about Call Analytics categories, see Creating categories for batch transcriptions and Creating categories for streaming transcriptions. To make a StartCallAnalyticsJob request, you must first upload your media file into an Amazon S3 bucket; you can then specify the Amazon S3 location of the file using the Media parameter. You must include the following parameters in your StartCallAnalyticsJob request: region: The Amazon Web Services Region where you are making your request. For a list of Amazon Web Services Regions supported with Amazon Transcribe, refer to Amazon Transcribe endpoints and quotas. CallAnalyticsJobName: A custom name that you create for your transcription job that's unique within your Amazon Web Services account. DataAccessRoleArn: The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. Media (MediaFileUri or RedactedMediaFileUri): The Amazon S3 location of your media file. With Call Analytics, you can redact the audio contained in your media file by including RedactedMediaFileUri, instead of MediaFileUri, to specify the location of your input audio. If you choose to redact your audio, you can find your redacted media at the location specified in the RedactedMediaFileUri field of your response.
260
260
  */
261
261
  startCallAnalyticsJob(callback?: (err: AWSError, data: TranscribeService.Types.StartCallAnalyticsJobResponse) => void): Request<TranscribeService.Types.StartCallAnalyticsJobResponse, AWSError>;
262
262
  /**
263
- * Transcribes the audio from a medical dictation or conversation and applies any additional Request Parameters you choose to include in your request. In addition to many of the standard transcription features, Amazon Transcribe Medical provides you with a robust medical vocabulary and, optionally, content identification, which adds flags to personal health information (PHI). To learn more about these features, refer to How Amazon Transcribe Medical works. To make a StartMedicalTranscriptionJob request, you must first upload your media file into an Amazon S3 bucket; you can then specify the S3 location of the file using the Media parameter. You must include the following parameters in your StartMedicalTranscriptionJob request: region: The Amazon Web Services Region where you are making your request. For a list of Amazon Web Services Regions supported with Amazon Transcribe, refer to Amazon Transcribe endpoints and quotas. MedicalTranscriptionJobName: A custom name you create for your transcription job that is unique within your Amazon Web Services account. Media (MediaFileUri): The Amazon S3 location of your media file. LanguageCode: This must be en-US. OutputBucketName: The Amazon S3 bucket where you want your transcript stored. If you want your output stored in a sub-folder of this bucket, you must also include OutputKey. Specialty: This must be PRIMARYCARE. Type: Choose whether your audio is a conversation or a dictation.
263
+ * Transcribes the audio from a medical dictation or conversation and applies any additional Request Parameters you choose to include in your request. In addition to many standard transcription features, Amazon Transcribe Medical provides you with a robust medical vocabulary and, optionally, content identification, which adds flags to personal health information (PHI). To learn more about these features, refer to How Amazon Transcribe Medical works. To make a StartMedicalTranscriptionJob request, you must first upload your media file into an Amazon S3 bucket; you can then specify the S3 location of the file using the Media parameter. You must include the following parameters in your StartMedicalTranscriptionJob request: region: The Amazon Web Services Region where you are making your request. For a list of Amazon Web Services Regions supported with Amazon Transcribe, refer to Amazon Transcribe endpoints and quotas. MedicalTranscriptionJobName: A custom name you create for your transcription job that is unique within your Amazon Web Services account. Media (MediaFileUri): The Amazon S3 location of your media file. LanguageCode: This must be en-US. OutputBucketName: The Amazon S3 bucket where you want your transcript stored. If you want your output stored in a sub-folder of this bucket, you must also include OutputKey. Specialty: This must be PRIMARYCARE. Type: Choose whether your audio is a conversation or a dictation.
264
264
  */
265
265
  startMedicalTranscriptionJob(params: TranscribeService.Types.StartMedicalTranscriptionJobRequest, callback?: (err: AWSError, data: TranscribeService.Types.StartMedicalTranscriptionJobResponse) => void): Request<TranscribeService.Types.StartMedicalTranscriptionJobResponse, AWSError>;
266
266
  /**
267
- * Transcribes the audio from a medical dictation or conversation and applies any additional Request Parameters you choose to include in your request. In addition to many of the standard transcription features, Amazon Transcribe Medical provides you with a robust medical vocabulary and, optionally, content identification, which adds flags to personal health information (PHI). To learn more about these features, refer to How Amazon Transcribe Medical works. To make a StartMedicalTranscriptionJob request, you must first upload your media file into an Amazon S3 bucket; you can then specify the S3 location of the file using the Media parameter. You must include the following parameters in your StartMedicalTranscriptionJob request: region: The Amazon Web Services Region where you are making your request. For a list of Amazon Web Services Regions supported with Amazon Transcribe, refer to Amazon Transcribe endpoints and quotas. MedicalTranscriptionJobName: A custom name you create for your transcription job that is unique within your Amazon Web Services account. Media (MediaFileUri): The Amazon S3 location of your media file. LanguageCode: This must be en-US. OutputBucketName: The Amazon S3 bucket where you want your transcript stored. If you want your output stored in a sub-folder of this bucket, you must also include OutputKey. Specialty: This must be PRIMARYCARE. Type: Choose whether your audio is a conversation or a dictation.
267
+ * Transcribes the audio from a medical dictation or conversation and applies any additional Request Parameters you choose to include in your request. In addition to many standard transcription features, Amazon Transcribe Medical provides you with a robust medical vocabulary and, optionally, content identification, which adds flags to personal health information (PHI). To learn more about these features, refer to How Amazon Transcribe Medical works. To make a StartMedicalTranscriptionJob request, you must first upload your media file into an Amazon S3 bucket; you can then specify the S3 location of the file using the Media parameter. You must include the following parameters in your StartMedicalTranscriptionJob request: region: The Amazon Web Services Region where you are making your request. For a list of Amazon Web Services Regions supported with Amazon Transcribe, refer to Amazon Transcribe endpoints and quotas. MedicalTranscriptionJobName: A custom name you create for your transcription job that is unique within your Amazon Web Services account. Media (MediaFileUri): The Amazon S3 location of your media file. LanguageCode: This must be en-US. OutputBucketName: The Amazon S3 bucket where you want your transcript stored. If you want your output stored in a sub-folder of this bucket, you must also include OutputKey. Specialty: This must be PRIMARYCARE. Type: Choose whether your audio is a conversation or a dictation.
268
268
  */
269
269
  startMedicalTranscriptionJob(callback?: (err: AWSError, data: TranscribeService.Types.StartMedicalTranscriptionJobResponse) => void): Request<TranscribeService.Types.StartMedicalTranscriptionJobResponse, AWSError>;
270
270
  /**
@@ -300,27 +300,27 @@ declare class TranscribeService extends Service {
300
300
  */
301
301
  updateCallAnalyticsCategory(callback?: (err: AWSError, data: TranscribeService.Types.UpdateCallAnalyticsCategoryResponse) => void): Request<TranscribeService.Types.UpdateCallAnalyticsCategoryResponse, AWSError>;
302
302
  /**
303
- * Updates an existing custom medical vocabulary with new values. This operation overwrites all existing information with your new values; you cannot append new terms onto an existing vocabulary.
303
+ * Updates an existing custom medical vocabulary with new values. This operation overwrites all existing information with your new values; you cannot append new terms onto an existing custom vocabulary.
304
304
  */
305
305
  updateMedicalVocabulary(params: TranscribeService.Types.UpdateMedicalVocabularyRequest, callback?: (err: AWSError, data: TranscribeService.Types.UpdateMedicalVocabularyResponse) => void): Request<TranscribeService.Types.UpdateMedicalVocabularyResponse, AWSError>;
306
306
  /**
307
- * Updates an existing custom medical vocabulary with new values. This operation overwrites all existing information with your new values; you cannot append new terms onto an existing vocabulary.
307
+ * Updates an existing custom medical vocabulary with new values. This operation overwrites all existing information with your new values; you cannot append new terms onto an existing custom vocabulary.
308
308
  */
309
309
  updateMedicalVocabulary(callback?: (err: AWSError, data: TranscribeService.Types.UpdateMedicalVocabularyResponse) => void): Request<TranscribeService.Types.UpdateMedicalVocabularyResponse, AWSError>;
310
310
  /**
311
- * Updates an existing custom vocabulary with new values. This operation overwrites all existing information with your new values; you cannot append new terms onto an existing vocabulary.
311
+ * Updates an existing custom vocabulary with new values. This operation overwrites all existing information with your new values; you cannot append new terms onto an existing custom vocabulary.
312
312
  */
313
313
  updateVocabulary(params: TranscribeService.Types.UpdateVocabularyRequest, callback?: (err: AWSError, data: TranscribeService.Types.UpdateVocabularyResponse) => void): Request<TranscribeService.Types.UpdateVocabularyResponse, AWSError>;
314
314
  /**
315
- * Updates an existing custom vocabulary with new values. This operation overwrites all existing information with your new values; you cannot append new terms onto an existing vocabulary.
315
+ * Updates an existing custom vocabulary with new values. This operation overwrites all existing information with your new values; you cannot append new terms onto an existing custom vocabulary.
316
316
  */
317
317
  updateVocabulary(callback?: (err: AWSError, data: TranscribeService.Types.UpdateVocabularyResponse) => void): Request<TranscribeService.Types.UpdateVocabularyResponse, AWSError>;
318
318
  /**
319
- * Updates an existing custom vocabulary filter with a new list of words. The new list you provide overwrites all previous entries; you cannot append new terms onto an existing vocabulary filter.
319
+ * Updates an existing custom vocabulary filter with a new list of words. The new list you provide overwrites all previous entries; you cannot append new terms onto an existing custom vocabulary filter.
320
320
  */
321
321
  updateVocabularyFilter(params: TranscribeService.Types.UpdateVocabularyFilterRequest, callback?: (err: AWSError, data: TranscribeService.Types.UpdateVocabularyFilterResponse) => void): Request<TranscribeService.Types.UpdateVocabularyFilterResponse, AWSError>;
322
322
  /**
323
- * Updates an existing custom vocabulary filter with a new list of words. The new list you provide overwrites all previous entries; you cannot append new terms onto an existing vocabulary filter.
323
+ * Updates an existing custom vocabulary filter with a new list of words. The new list you provide overwrites all previous entries; you cannot append new terms onto an existing custom vocabulary filter.
324
324
  */
325
325
  updateVocabularyFilter(callback?: (err: AWSError, data: TranscribeService.Types.UpdateVocabularyFilterResponse) => void): Request<TranscribeService.Types.UpdateVocabularyFilterResponse, AWSError>;
326
326
  }
@@ -335,11 +335,11 @@ declare namespace TranscribeService {
335
335
  */
336
336
  EndTime?: TimestampMilliseconds;
337
337
  /**
338
- * The time, in milliseconds, from the start of your media file until the value you specify in which Amazon Transcribe searches for your specified criteria.
338
+ * The time, in milliseconds, from the start of your media file until the specified value. Amazon Transcribe searches for your specified criteria in this time segment.
339
339
  */
340
340
  First?: TimestampMilliseconds;
341
341
  /**
342
- * The time, in milliseconds, from the value you specify until the end of your media file in which Amazon Transcribe searches for your specified criteria.
342
+ * The time, in milliseconds, from the specified value until the end of your media file. Amazon Transcribe searches for your specified criteria in this time segment.
343
343
  */
344
344
  Last?: TimestampMilliseconds;
345
345
  }
@@ -360,13 +360,16 @@ declare namespace TranscribeService {
360
360
  */
361
361
  LanguageCode?: LanguageCode;
362
362
  /**
363
- * The sample rate, in Hertz, of the audio track in your input media file.
363
+ * The sample rate, in hertz, of the audio track in your input media file.
364
364
  */
365
365
  MediaSampleRateHertz?: MediaSampleRateHertz;
366
366
  /**
367
367
  * The format of the input media file.
368
368
  */
369
369
  MediaFormat?: MediaFormat;
370
+ /**
371
+ * Provides the Amazon S3 location of the media file you used in your Call Analytics request.
372
+ */
370
373
  Media?: Media;
371
374
  Transcript?: Transcript;
372
375
  /**
@@ -382,11 +385,11 @@ declare namespace TranscribeService {
382
385
  */
383
386
  CompletionTime?: DateTime;
384
387
  /**
385
- * If CallAnalyticsJobStatus is FAILED, FailureReason contains information about why the Call Analytics job request failed. The FailureReason field contains one of the following values: Unsupported media format. The media format specified in MediaFormat isn't valid. Refer to MediaFormat for a list of supported formats. The media format provided does not match the detected media format. The media format specified in MediaFormat doesn't match the format of the input file. Check the media format of your media file and correct the specified value. Invalid sample rate for audio file. The sample rate specified in MediaSampleRateHertz isn't valid. The sample rate must be between 8,000 and 48,000 Hertz. The sample rate provided does not match the detected sample rate. The sample rate specified in MediaSampleRateHertz doesn't match the sample rate detected in your input media file. Check the sample rate of your media file and correct the specified value. Invalid file size: file size too large. The size of your media file is larger than what Amazon Transcribe can process. For more information, refer to Guidelines and quotas. Invalid number of channels: number of channels too large. Your audio contains more channels than Amazon Transcribe is able to process. For more information, refer to Guidelines and quotas.
388
+ * If CallAnalyticsJobStatus is FAILED, FailureReason contains information about why the Call Analytics job request failed. The FailureReason field contains one of the following values: Unsupported media format. The media format specified in MediaFormat isn't valid. Refer to MediaFormat for a list of supported formats. The media format provided does not match the detected media format. The media format specified in MediaFormat doesn't match the format of the input file. Check the media format of your media file and correct the specified value. Invalid sample rate for audio file. The sample rate specified in MediaSampleRateHertz isn't valid. The sample rate must be between 8,000 and 48,000 hertz. The sample rate provided does not match the detected sample rate. The sample rate specified in MediaSampleRateHertz doesn't match the sample rate detected in your input media file. Check the sample rate of your media file and correct the specified value. Invalid file size: file size too large. The size of your media file is larger than what Amazon Transcribe can process. For more information, refer to Guidelines and quotas. Invalid number of channels: number of channels too large. Your audio contains more channels than Amazon Transcribe is able to process. For more information, refer to Guidelines and quotas.
386
389
  */
387
390
  FailureReason?: FailureReason;
388
391
  /**
389
- * The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. If the role you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails. IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path. For example: arn:aws:iam::111122223333:role/Admin. For more information, see IAM ARNs.
392
+ * The Amazon Resource Name (ARN) you included in your request.
390
393
  */
391
394
  DataAccessRoleArn?: DataAccessRoleArn;
392
395
  /**
@@ -394,39 +397,39 @@ declare namespace TranscribeService {
394
397
  */
395
398
  IdentifiedLanguageScore?: IdentifiedLanguageScore;
396
399
  /**
397
- * Allows additional optional settings in your request, including content redaction; allows you to apply custom language models, vocabulary filters, and custom vocabularies to your Call Analytics job.
400
+ * Provides information on any additional settings that were included in your request. Additional settings include content redaction and language identification settings.
398
401
  */
399
402
  Settings?: CallAnalyticsJobSettings;
400
403
  /**
401
- * Allows you to specify which speaker is on which channel in your Call Analytics job request. For example, if your agent is the first participant to speak, you would set ChannelId to 0 (to indicate the first channel) and ParticipantRole to AGENT (to indicate that it's the agent speaking).
404
+ * Indicates which speaker is on which channel.
402
405
  */
403
406
  ChannelDefinitions?: ChannelDefinitions;
404
407
  }
405
408
  export type CallAnalyticsJobName = string;
406
409
  export interface CallAnalyticsJobSettings {
407
410
  /**
408
- * The name of the custom vocabulary you want to include in your Call Analytics transcription request. Vocabulary names are case sensitive.
411
+ * The name of the custom vocabulary you want to include in your Call Analytics transcription request. Custom vocabulary names are case sensitive.
409
412
  */
410
413
  VocabularyName?: VocabularyName;
411
414
  /**
412
- * The name of the custom vocabulary filter you want to include in your Call Analytics transcription request. Vocabulary filter names are case sensitive. Note that if you include VocabularyFilterName in your request, you must also include VocabularyFilterMethod.
415
+ * The name of the custom vocabulary filter you want to include in your Call Analytics transcription request. Custom vocabulary filter names are case sensitive. Note that if you include VocabularyFilterName in your request, you must also include VocabularyFilterMethod.
413
416
  */
414
417
  VocabularyFilterName?: VocabularyFilterName;
415
418
  /**
416
- * Specify how you want your vocabulary filter applied to your transcript. To replace words with ***, choose mask. To delete words, choose remove. To flag words without changing them, choose tag.
419
+ * Specify how you want your custom vocabulary filter applied to your transcript. To replace words with ***, choose mask. To delete words, choose remove. To flag words without changing them, choose tag.
417
420
  */
418
421
  VocabularyFilterMethod?: VocabularyFilterMethod;
419
422
  /**
420
- * The name of the custom language model you want to use when processing your Call Analytics job. Note that language model names are case sensitive. The language of the specified language model must match the language code you specify in your transcription request. If the languages don't match, the language model isn't applied. There are no errors or warnings associated with a language mismatch.
423
+ * The name of the custom language model you want to use when processing your Call Analytics job. Note that custom language model names are case sensitive. The language of the specified custom language model must match the language code that you specify in your transcription request. If the languages don't match, the custom language model isn't applied. There are no errors or warnings associated with a language mismatch.
421
424
  */
422
425
  LanguageModelName?: ModelName;
423
426
  ContentRedaction?: ContentRedaction;
424
427
  /**
425
- * You can specify two or more language codes that represent the languages you think may be present in your media; including more than five is not recommended. If you're unsure what languages are present, do not include this parameter. Including language options can improve the accuracy of language identification. For a list of languages supported with Call Analytics, refer to the Supported languages table.
428
+ * You can specify two or more language codes that represent the languages you think may be present in your media. Including more than five is not recommended. If you're unsure what languages are present, do not include this parameter. Including language options can improve the accuracy of language identification. For a list of languages supported with Call Analytics, refer to the Supported languages table. To transcribe speech in Modern Standard Arabic (ar-SA), your media file must be encoded at a sample rate of 16,000 Hz or higher.
426
429
  */
427
430
  LanguageOptions?: LanguageOptions;
428
431
  /**
429
- * If using automatic language identification (IdentifyLanguage) in your request and you want to apply a custom language model, a custom vocabulary, or a custom vocabulary filter, include LanguageIdSettings with the relevant sub-parameters (VocabularyName, LanguageModelName, and VocabularyFilterName). You can specify two or more language codes that represent the languages you think may be present in your media; including more than five is not recommended. Each language code you include can have an associated custom language model, custom vocabulary, and custom vocabulary filter. The languages you specify must match the languages of the specified custom language models, custom vocabularies, and custom vocabulary filters. To include language options using IdentifyLanguage without including a custom language model, a custom vocabulary, or a custom vocabulary filter, use LanguageOptions instead of LanguageIdSettings. Including language options can improve the accuracy of automatic language identification. If you want to include a custom language model with your request but do not want to use automatic language identification, use instead the parameter with the LanguageModelName sub-parameter. If you want to include a custom vocabulary or a custom vocabulary filter (or both) with your request but do not want to use automatic language identification, use instead the parameter with the VocabularyName or VocabularyFilterName (or both) sub-parameter.
432
+ * If using automatic language identification in your request and you want to apply a custom language model, a custom vocabulary, or a custom vocabulary filter, include LanguageIdSettings with the relevant sub-parameters (VocabularyName, LanguageModelName, and VocabularyFilterName). LanguageIdSettings supports two to five language codes. Each language code you include can have an associated custom language model, custom vocabulary, and custom vocabulary filter. The language codes that you specify must match the languages of the associated custom language models, custom vocabularies, and custom vocabulary filters. It's recommended that you include LanguageOptions when using LanguageIdSettings to ensure that the correct language dialect is identified. For example, if you specify a custom vocabulary that is in en-US but Amazon Transcribe determines that the language spoken in your media is en-AU, your custom vocabulary is not applied to your transcription. If you include LanguageOptions and include en-US as the only English language dialect, your custom vocabulary is applied to your transcription. If you want to include a custom language model, custom vocabulary, or custom vocabulary filter with your request but do not want to use automatic language identification, use instead the parameter with the LanguageModelName, VocabularyName, or VocabularyFilterName sub-parameters. For a list of languages supported with Call Analytics, refer to Supported languages and language-specific features.
430
433
  */
431
434
  LanguageIdSettings?: LanguageIdSettingsMap;
432
435
  }
@@ -480,6 +483,10 @@ declare namespace TranscribeService {
480
483
  * The date and time the specified Call Analytics category was last updated. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-05T12:45:32.691000-07:00 represents 12:45 PM UTC-7 on May 5, 2022.
481
484
  */
482
485
  LastUpdateTime?: DateTime;
486
+ /**
487
+ * The input type associated with the specified category. POST_CALL refers to a category that is applied to batch transcriptions; REAL_TIME refers to a category that is applied to streaming transcriptions.
488
+ */
489
+ InputType?: InputType;
483
490
  }
484
491
  export type CategoryPropertiesList = CategoryProperties[];
485
492
  export interface ChannelDefinition {
@@ -514,9 +521,13 @@ declare namespace TranscribeService {
514
521
  */
515
522
  CategoryName: CategoryName;
516
523
  /**
517
- * Rules define a Call Analytics category. When creating a new Call Analytics category, you must create between 1 and 20 rules for that category. For each rule, you specify a filter you want applied to the attributes of a call. For example, you can choose a sentiment filter that detects if a customer's sentiment was positive during the last 30 seconds of the call.
524
+ * Rules define a Call Analytics category. When creating a new category, you must create between 1 and 20 rules for that category. For each rule, you specify a filter you want applied to the attributes of a call. For example, you can choose a sentiment filter that detects if a customer's sentiment was positive during the last 30 seconds of the call.
518
525
  */
519
526
  Rules: RuleList;
527
+ /**
528
+ * Choose whether you want to create a streaming or a batch category for your Call Analytics transcription. Specifying POST_CALL assigns your category to batch transcriptions; categories with this input type cannot be applied to streaming (real-time) transcriptions. Specifying REAL_TIME assigns your category to streaming transcriptions; categories with this input type cannot be applied to batch (post-call) transcriptions. If you do not include InputType, your category is created as a batch category by default.
529
+ */
530
+ InputType?: InputType;
520
531
  }
521
532
  export interface CreateCallAnalyticsCategoryResponse {
522
533
  /**
@@ -526,7 +537,7 @@ declare namespace TranscribeService {
526
537
  }
527
538
  export interface CreateLanguageModelRequest {
528
539
  /**
529
- * The language code that represents the language of your model. Each language model must contain terms in only one language, and the language you select for your model must match the language of your training and tuning data. For a list of supported languages and their associated language codes, refer to the Supported languages table. Note that U.S. English (en-US) is the only language supported with Amazon Transcribe Medical. A custom language model can only be used to transcribe files in the same language as the model. For example, if you create a language model using US English (en-US), you can only apply this model to files that contain English audio.
540
+ * The language code that represents the language of your model. Each custom language model must contain terms in only one language, and the language you select for your custom language model must match the language of your training and tuning data. For a list of supported languages and their associated language codes, refer to the Supported languages table. Note that US English (en-US) is the only language supported with Amazon Transcribe Medical. A custom language model can only be used to transcribe files in the same language as the model. For example, if you create a custom language model using US English (en-US), you can only apply this model to files that contain English audio.
530
541
  */
531
542
  LanguageCode: CLMLanguageCode;
532
543
  /**
@@ -534,7 +545,7 @@ declare namespace TranscribeService {
534
545
  */
535
546
  BaseModelName: BaseModelName;
536
547
  /**
537
- * A unique name, chosen by you, for your custom language model. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new language model with the same name as an existing language model, you get a ConflictException error.
548
+ * A unique name, chosen by you, for your custom language model. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new custom language model with the same name as an existing custom language model, you get a ConflictException error.
538
549
  */
539
550
  ModelName: ModelName;
540
551
  /**
@@ -570,7 +581,7 @@ declare namespace TranscribeService {
570
581
  }
571
582
  export interface CreateMedicalVocabularyRequest {
572
583
  /**
573
- * A unique name, chosen by you, for your new custom medical vocabulary. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new medical vocabulary with the same name as an existing medical vocabulary, you get a ConflictException error.
584
+ * A unique name, chosen by you, for your new custom medical vocabulary. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new custom medical vocabulary with the same name as an existing custom medical vocabulary, you get a ConflictException error.
574
585
  */
575
586
  VocabularyName: VocabularyName;
576
587
  /**
@@ -582,7 +593,7 @@ declare namespace TranscribeService {
582
593
  */
583
594
  VocabularyFileUri: Uri;
584
595
  /**
585
- * Adds one or more custom tags, each in the form of a key:value pair, to a new medical vocabulary at the time you create this new vocabulary. To learn more about using tags with Amazon Transcribe, refer to Tagging resources.
596
+ * Adds one or more custom tags, each in the form of a key:value pair, to a new custom medical vocabulary at the time you create this new custom vocabulary. To learn more about using tags with Amazon Transcribe, refer to Tagging resources.
586
597
  */
587
598
  Tags?: TagList;
588
599
  }
@@ -592,11 +603,11 @@ declare namespace TranscribeService {
592
603
  */
593
604
  VocabularyName?: VocabularyName;
594
605
  /**
595
- * The language code you selected for your medical vocabulary. US English (en-US) is the only language supported with Amazon Transcribe Medical.
606
+ * The language code you selected for your custom medical vocabulary. US English (en-US) is the only language supported with Amazon Transcribe Medical.
596
607
  */
597
608
  LanguageCode?: LanguageCode;
598
609
  /**
599
- * The processing state of your custom medical vocabulary. If the state is READY, you can use the vocabulary in a StartMedicalTranscriptionJob request.
610
+ * The processing state of your custom medical vocabulary. If the state is READY, you can use the custom vocabulary in a StartMedicalTranscriptionJob request.
600
611
  */
601
612
  VocabularyState?: VocabularyState;
602
613
  /**
@@ -610,15 +621,15 @@ declare namespace TranscribeService {
610
621
  }
611
622
  export interface CreateVocabularyFilterRequest {
612
623
  /**
613
- * A unique name, chosen by you, for your new custom vocabulary filter. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new vocabulary filter with the same name as an existing vocabulary filter, you get a ConflictException error.
624
+ * A unique name, chosen by you, for your new custom vocabulary filter. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new custom vocabulary filter with the same name as an existing custom vocabulary filter, you get a ConflictException error.
614
625
  */
615
626
  VocabularyFilterName: VocabularyFilterName;
616
627
  /**
617
- * The language code that represents the language of the entries in your vocabulary filter. Each vocabulary filter must contain terms in only one language. A vocabulary filter can only be used to transcribe files in the same language as the filter. For example, if you create a vocabulary filter using US English (en-US), you can only apply this filter to files that contain English audio. For a list of supported languages and their associated language codes, refer to the Supported languages table.
628
+ * The language code that represents the language of the entries in your vocabulary filter. Each custom vocabulary filter must contain terms in only one language. A custom vocabulary filter can only be used to transcribe files in the same language as the filter. For example, if you create a custom vocabulary filter using US English (en-US), you can only apply this filter to files that contain English audio. For a list of supported languages and their associated language codes, refer to the Supported languages table.
618
629
  */
619
630
  LanguageCode: LanguageCode;
620
631
  /**
621
- * Use this parameter if you want to create your vocabulary filter by including all desired terms, as comma-separated values, within your request. The other option for creating your vocabulary filter is to save your entries in a text file and upload them to an Amazon S3 bucket, then specify the location of your file using the VocabularyFilterFileUri parameter. Note that if you include Words in your request, you cannot use VocabularyFilterFileUri; you must choose one or the other. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language.
632
+ * Use this parameter if you want to create your custom vocabulary filter by including all desired terms, as comma-separated values, within your request. The other option for creating your vocabulary filter is to save your entries in a text file and upload them to an Amazon S3 bucket, then specify the location of your file using the VocabularyFilterFileUri parameter. Note that if you include Words in your request, you cannot use VocabularyFilterFileUri; you must choose one or the other. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language.
622
633
  */
623
634
  Words?: Words;
624
635
  /**
@@ -626,7 +637,7 @@ declare namespace TranscribeService {
626
637
  */
627
638
  VocabularyFilterFileUri?: Uri;
628
639
  /**
629
- * Adds one or more custom tags, each in the form of a key:value pair, to a new custom vocabulary filter at the time you create this new filter. To learn more about using tags with Amazon Transcribe, refer to Tagging resources.
640
+ * Adds one or more custom tags, each in the form of a key:value pair, to a new custom vocabulary filter at the time you create this new vocabulary filter. To learn more about using tags with Amazon Transcribe, refer to Tagging resources.
630
641
  */
631
642
  Tags?: TagList;
632
643
  }
@@ -636,25 +647,25 @@ declare namespace TranscribeService {
636
647
  */
637
648
  VocabularyFilterName?: VocabularyFilterName;
638
649
  /**
639
- * The language code you selected for your vocabulary filter.
650
+ * The language code you selected for your custom vocabulary filter.
640
651
  */
641
652
  LanguageCode?: LanguageCode;
642
653
  /**
643
- * The date and time you created your vocabulary filter. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
654
+ * The date and time you created your custom vocabulary filter. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
644
655
  */
645
656
  LastModifiedTime?: DateTime;
646
657
  }
647
658
  export interface CreateVocabularyRequest {
648
659
  /**
649
- * A unique name, chosen by you, for your new custom vocabulary. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new vocabulary with the same name as an existing vocabulary, you get a ConflictException error.
660
+ * A unique name, chosen by you, for your new custom vocabulary. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new custom vocabulary with the same name as an existing custom vocabulary, you get a ConflictException error.
650
661
  */
651
662
  VocabularyName: VocabularyName;
652
663
  /**
653
- * The language code that represents the language of the entries in your custom vocabulary. Each vocabulary must contain terms in only one language. A custom vocabulary can only be used to transcribe files in the same language as the vocabulary. For example, if you create a vocabulary using US English (en-US), you can only apply this vocabulary to files that contain English audio. For a list of supported languages and their associated language codes, refer to the Supported languages table.
664
+ * The language code that represents the language of the entries in your custom vocabulary. Each custom vocabulary must contain terms in only one language. A custom vocabulary can only be used to transcribe files in the same language as the custom vocabulary. For example, if you create a custom vocabulary using US English (en-US), you can only apply this custom vocabulary to files that contain English audio. For a list of supported languages and their associated language codes, refer to the Supported languages table.
654
665
  */
655
666
  LanguageCode: LanguageCode;
656
667
  /**
657
- * Use this parameter if you want to create your vocabulary by including all desired terms, as comma-separated values, within your request. The other option for creating your vocabulary is to save your entries in a text file and upload them to an Amazon S3 bucket, then specify the location of your file using the VocabularyFileUri parameter. Note that if you include Phrases in your request, you cannot use VocabularyFileUri; you must choose one or the other. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language.
668
+ * Use this parameter if you want to create your custom vocabulary by including all desired terms, as comma-separated values, within your request. The other option for creating your custom vocabulary is to save your entries in a text file and upload them to an Amazon S3 bucket, then specify the location of your file using the VocabularyFileUri parameter. Note that if you include Phrases in your request, you cannot use VocabularyFileUri; you must choose one or the other. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language.
658
669
  */
659
670
  Phrases?: Phrases;
660
671
  /**
@@ -662,7 +673,7 @@ declare namespace TranscribeService {
662
673
  */
663
674
  VocabularyFileUri?: Uri;
664
675
  /**
665
- * Adds one or more custom tags, each in the form of a key:value pair, to a new custom vocabulary at the time you create this new vocabulary. To learn more about using tags with Amazon Transcribe, refer to Tagging resources.
676
+ * Adds one or more custom tags, each in the form of a key:value pair, to a new custom vocabulary at the time you create this new custom vocabulary. To learn more about using tags with Amazon Transcribe, refer to Tagging resources.
666
677
  */
667
678
  Tags?: TagList;
668
679
  }
@@ -676,7 +687,7 @@ declare namespace TranscribeService {
676
687
  */
677
688
  LanguageCode?: LanguageCode;
678
689
  /**
679
- * The processing state of your custom vocabulary. If the state is READY, you can use the vocabulary in a StartTranscriptionJob request.
690
+ * The processing state of your custom vocabulary. If the state is READY, you can use the custom vocabulary in a StartTranscriptionJob request.
680
691
  */
681
692
  VocabularyState?: VocabularyState;
682
693
  /**
@@ -684,7 +695,7 @@ declare namespace TranscribeService {
684
695
  */
685
696
  LastModifiedTime?: DateTime;
686
697
  /**
687
- * If VocabularyState is FAILED, FailureReason contains information about why the vocabulary request failed. See also: Common Errors.
698
+ * If VocabularyState is FAILED, FailureReason contains information about why the custom vocabulary request failed. See also: Common Errors.
688
699
  */
689
700
  FailureReason?: FailureReason;
690
701
  }
@@ -720,7 +731,7 @@ declare namespace TranscribeService {
720
731
  }
721
732
  export interface DeleteMedicalVocabularyRequest {
722
733
  /**
723
- * The name of the custom medical vocabulary you want to delete. Vocabulary names are case sensitive.
734
+ * The name of the custom medical vocabulary you want to delete. Custom medical vocabulary names are case sensitive.
724
735
  */
725
736
  VocabularyName: VocabularyName;
726
737
  }
@@ -732,13 +743,13 @@ declare namespace TranscribeService {
732
743
  }
733
744
  export interface DeleteVocabularyFilterRequest {
734
745
  /**
735
- * The name of the custom vocabulary filter you want to delete. Vocabulary filter names are case sensitive.
746
+ * The name of the custom vocabulary filter you want to delete. Custom vocabulary filter names are case sensitive.
736
747
  */
737
748
  VocabularyFilterName: VocabularyFilterName;
738
749
  }
739
750
  export interface DeleteVocabularyRequest {
740
751
  /**
741
- * The name of the custom vocabulary you want to delete. Vocabulary names are case sensitive.
752
+ * The name of the custom vocabulary you want to delete. Custom vocabulary names are case sensitive.
742
753
  */
743
754
  VocabularyName: VocabularyName;
744
755
  }
@@ -794,7 +805,7 @@ declare namespace TranscribeService {
794
805
  }
795
806
  export interface GetMedicalVocabularyRequest {
796
807
  /**
797
- * The name of the custom medical vocabulary you want information about. Vocabulary names are case sensitive.
808
+ * The name of the custom medical vocabulary you want information about. Custom medical vocabulary names are case sensitive.
798
809
  */
799
810
  VocabularyName: VocabularyName;
800
811
  }
@@ -804,11 +815,11 @@ declare namespace TranscribeService {
804
815
  */
805
816
  VocabularyName?: VocabularyName;
806
817
  /**
807
- * The language code you selected for your medical vocabulary. US English (en-US) is the only language supported with Amazon Transcribe Medical.
818
+ * The language code you selected for your custom medical vocabulary. US English (en-US) is the only language supported with Amazon Transcribe Medical.
808
819
  */
809
820
  LanguageCode?: LanguageCode;
810
821
  /**
811
- * The processing state of your custom medical vocabulary. If the state is READY, you can use the vocabulary in a StartMedicalTranscriptionJob request.
822
+ * The processing state of your custom medical vocabulary. If the state is READY, you can use the custom vocabulary in a StartMedicalTranscriptionJob request.
812
823
  */
813
824
  VocabularyState?: VocabularyState;
814
825
  /**
@@ -816,11 +827,11 @@ declare namespace TranscribeService {
816
827
  */
817
828
  LastModifiedTime?: DateTime;
818
829
  /**
819
- * If VocabularyState is FAILED, FailureReason contains information about why the medical vocabulary request failed. See also: Common Errors.
830
+ * If VocabularyState is FAILED, FailureReason contains information about why the custom medical vocabulary request failed. See also: Common Errors.
820
831
  */
821
832
  FailureReason?: FailureReason;
822
833
  /**
823
- * The S3 location where the specified medical vocabulary is stored; use this URI to view or download the vocabulary.
834
+ * The S3 location where the specified custom medical vocabulary is stored; use this URI to view or download the custom vocabulary.
824
835
  */
825
836
  DownloadUri?: Uri;
826
837
  }
@@ -838,7 +849,7 @@ declare namespace TranscribeService {
838
849
  }
839
850
  export interface GetVocabularyFilterRequest {
840
851
  /**
841
- * The name of the custom vocabulary filter you want information about. Vocabulary filter names are case sensitive.
852
+ * The name of the custom vocabulary filter you want information about. Custom vocabulary filter names are case sensitive.
842
853
  */
843
854
  VocabularyFilterName: VocabularyFilterName;
844
855
  }
@@ -848,21 +859,21 @@ declare namespace TranscribeService {
848
859
  */
849
860
  VocabularyFilterName?: VocabularyFilterName;
850
861
  /**
851
- * The language code you selected for your vocabulary filter.
862
+ * The language code you selected for your custom vocabulary filter.
852
863
  */
853
864
  LanguageCode?: LanguageCode;
854
865
  /**
855
- * The date and time the specified vocabulary filter was last modified. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
866
+ * The date and time the specified custom vocabulary filter was last modified. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
856
867
  */
857
868
  LastModifiedTime?: DateTime;
858
869
  /**
859
- * The Amazon S3 location where the vocabulary filter is stored; use this URI to view or download the vocabulary filter.
870
+ * The Amazon S3 location where the custom vocabulary filter is stored; use this URI to view or download the custom vocabulary filter.
860
871
  */
861
872
  DownloadUri?: Uri;
862
873
  }
863
874
  export interface GetVocabularyRequest {
864
875
  /**
865
- * The name of the custom vocabulary you want information about. Vocabulary names are case sensitive.
876
+ * The name of the custom vocabulary you want information about. Custom vocabulary names are case sensitive.
866
877
  */
867
878
  VocabularyName: VocabularyName;
868
879
  }
@@ -876,19 +887,19 @@ declare namespace TranscribeService {
876
887
  */
877
888
  LanguageCode?: LanguageCode;
878
889
  /**
879
- * The processing state of your custom vocabulary. If the state is READY, you can use the vocabulary in a StartTranscriptionJob request.
890
+ * The processing state of your custom vocabulary. If the state is READY, you can use the custom vocabulary in a StartTranscriptionJob request.
880
891
  */
881
892
  VocabularyState?: VocabularyState;
882
893
  /**
883
- * The date and time the specified vocabulary was last modified. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
894
+ * The date and time the specified custom vocabulary was last modified. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
884
895
  */
885
896
  LastModifiedTime?: DateTime;
886
897
  /**
887
- * If VocabularyState is FAILED, FailureReason contains information about why the vocabulary request failed. See also: Common Errors.
898
+ * If VocabularyState is FAILED, FailureReason contains information about why the custom vocabulary request failed. See also: Common Errors.
888
899
  */
889
900
  FailureReason?: FailureReason;
890
901
  /**
891
- * The S3 location where the vocabulary is stored; use this URI to view or download the vocabulary.
902
+ * The S3 location where the custom vocabulary is stored; use this URI to view or download the custom vocabulary.
892
903
  */
893
904
  DownloadUri?: Uri;
894
905
  }
@@ -903,25 +914,26 @@ declare namespace TranscribeService {
903
914
  */
904
915
  TuningDataS3Uri?: Uri;
905
916
  /**
906
- * The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. If the role you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails. IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path. For example: arn:aws:iam::111122223333:role/Admin. For more information, see IAM ARNs.
917
+ * The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. If the role that you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails. IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path. For example: arn:aws:iam::111122223333:role/Admin. For more information, see IAM ARNs.
907
918
  */
908
919
  DataAccessRoleArn: DataAccessRoleArn;
909
920
  }
921
+ export type InputType = "REAL_TIME"|"POST_CALL"|string;
910
922
  export interface InterruptionFilter {
911
923
  /**
912
- * Specify the duration of the interruptions in milliseconds. For example, you can flag speech that contains more than 10000 milliseconds of interruptions.
924
+ * Specify the duration of the interruptions in milliseconds. For example, you can flag speech that contains more than 10,000 milliseconds of interruptions.
913
925
  */
914
926
  Threshold?: TimestampMilliseconds;
915
927
  /**
916
- * Specify the interrupter you want to flag. Omitting this parameter is equivalent to specifying both participants.
928
+ * Specify the interrupter that you want to flag. Omitting this parameter is equivalent to specifying both participants.
917
929
  */
918
930
  ParticipantRole?: ParticipantRole;
919
931
  /**
920
- * Allows you to specify a time range (in milliseconds) in your audio, during which you want to search for an interruption. See for more detail.
932
+ * Makes it possible to specify a time range (in milliseconds) in your audio, during which you want to search for an interruption. See for more detail.
921
933
  */
922
934
  AbsoluteTimeRange?: AbsoluteTimeRange;
923
935
  /**
924
- * Allows you to specify a time range (in percentage) in your media file, during which you want to search for an interruption. See for more detail.
936
+ * Makes it possible to specify a time range (in percentage) in your media file, during which you want to search for an interruption. See for more detail.
925
937
  */
926
938
  RelativeTimeRange?: RelativeTimeRange;
927
939
  /**
@@ -931,11 +943,11 @@ declare namespace TranscribeService {
931
943
  }
932
944
  export interface JobExecutionSettings {
933
945
  /**
934
- * Allows you to enable job queuing when your concurrent request limit is exceeded. When AllowDeferredExecution is set to true, transcription job requests are placed in a queue until the number of jobs falls below the concurrent request limit. If AllowDeferredExecution is set to false and the number of transcription job requests exceed the concurrent request limit, you get a LimitExceededException error. Note that job queuing is enabled by default for Call Analytics jobs. If you include AllowDeferredExecution in your request, you must also include DataAccessRoleArn.
946
+ * Makes it possible to enable job queuing when your concurrent request limit is exceeded. When AllowDeferredExecution is set to true, transcription job requests are placed in a queue until the number of jobs falls below the concurrent request limit. If AllowDeferredExecution is set to false and the number of transcription job requests exceed the concurrent request limit, you get a LimitExceededException error. Note that job queuing is enabled by default for Call Analytics jobs. If you include AllowDeferredExecution in your request, you must also include DataAccessRoleArn.
935
947
  */
936
948
  AllowDeferredExecution?: Boolean;
937
949
  /**
938
- * The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. If the role you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails. IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path. For example: arn:aws:iam::111122223333:role/Admin. For more information, see IAM ARNs. Note that if you include DataAccessRoleArn in your request, you must also include AllowDeferredExecution.
950
+ * The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. If the role that you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails. IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path. For example: arn:aws:iam::111122223333:role/Admin. For more information, see IAM ARNs. Note that if you include DataAccessRoleArn in your request, you must also include AllowDeferredExecution.
939
951
  */
940
952
  DataAccessRoleArn?: DataAccessRoleArn;
941
953
  }
@@ -955,15 +967,15 @@ declare namespace TranscribeService {
955
967
  export type LanguageCodeList = LanguageCodeItem[];
956
968
  export interface LanguageIdSettings {
957
969
  /**
958
- * The name of the custom vocabulary you want to use when processing your transcription job. Vocabulary names are case sensitive. The language of the specified vocabulary must match the language code you specify in your transcription request. If the languages don't match, the vocabulary isn't applied. There are no errors or warnings associated with a language mismatch.
970
+ * The name of the custom vocabulary you want to use when processing your transcription job. Custom vocabulary names are case sensitive. The language of the specified custom vocabulary must match the language code that you specify in your transcription request. If the languages don't match, the custom vocabulary isn't applied. There are no errors or warnings associated with a language mismatch.
959
971
  */
960
972
  VocabularyName?: VocabularyName;
961
973
  /**
962
- * The name of the custom vocabulary filter you want to use when processing your transcription job. Vocabulary filter names are case sensitive. The language of the specified vocabulary filter must match the language code you specify in your transcription request. If the languages don't match, the vocabulary filter isn't applied. There are no errors or warnings associated with a language mismatch. Note that if you include VocabularyFilterName in your request, you must also include VocabularyFilterMethod.
974
+ * The name of the custom vocabulary filter you want to use when processing your transcription job. Custom vocabulary filter names are case sensitive. The language of the specified custom vocabulary filter must match the language code that you specify in your transcription request. If the languages don't match, the custom vocabulary filter isn't applied. There are no errors or warnings associated with a language mismatch. Note that if you include VocabularyFilterName in your request, you must also include VocabularyFilterMethod.
963
975
  */
964
976
  VocabularyFilterName?: VocabularyFilterName;
965
977
  /**
966
- * The name of the custom language model you want to use when processing your transcription job. Note that language model names are case sensitive. The language of the specified language model must match the language code you specify in your transcription request. If the languages don't match, the language model isn't applied. There are no errors or warnings associated with a language mismatch.
978
+ * The name of the custom language model you want to use when processing your transcription job. Note that custom language model names are case sensitive. The language of the specified custom language model must match the language code that you specify in your transcription request. If the languages don't match, the custom language model isn't applied. There are no errors or warnings associated with a language mismatch.
967
979
  */
968
980
  LanguageModelName?: ModelName;
969
981
  }
@@ -978,11 +990,11 @@ declare namespace TranscribeService {
978
990
  */
979
991
  CreateTime?: DateTime;
980
992
  /**
981
- * The date and time the specified language model was last modified. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
993
+ * The date and time the specified custom language model was last modified. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
982
994
  */
983
995
  LastModifiedTime?: DateTime;
984
996
  /**
985
- * The language code used to create your custom language model. Each language model must contain terms in only one language, and the language you select for your model must match the language of your training and tuning data. For a list of supported languages and their associated language codes, refer to the Supported languages table. Note that U.S. English (en-US) is the only language supported with Amazon Transcribe Medical.
997
+ * The language code used to create your custom language model. Each custom language model must contain terms in only one language, and the language you select for your custom language model must match the language of your training and tuning data. For a list of supported languages and their associated language codes, refer to the Supported languages table. Note that U.S. English (en-US) is the only language supported with Amazon Transcribe Medical.
986
998
  */
987
999
  LanguageCode?: CLMLanguageCode;
988
1000
  /**
@@ -994,7 +1006,7 @@ declare namespace TranscribeService {
994
1006
  */
995
1007
  ModelStatus?: ModelStatus;
996
1008
  /**
997
- * Shows if a more current base model is available for use with the specified custom language model. If false, your language model is using the most up-to-date base model. If true, there is a newer base model available than the one your language model is using. Note that to update a base model, you must recreate the custom language model using the new base model. Base model upgrades for existing custom language models are not supported.
1009
+ * Shows if a more current base model is available for use with the specified custom language model. If false, your custom language model is using the most up-to-date base model. If true, there is a newer base model available than the one your language model is using. Note that to update a base model, you must recreate the custom language model using the new base model. Base model upgrades for existing custom language models are not supported.
998
1010
  */
999
1011
  UpgradeAvailability?: Boolean;
1000
1012
  /**
@@ -1013,7 +1025,7 @@ declare namespace TranscribeService {
1013
1025
  */
1014
1026
  NextToken?: NextToken;
1015
1027
  /**
1016
- * The maximum number of Call Analytics categories to return in each page of results. If there are fewer results than the value you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1028
+ * The maximum number of Call Analytics categories to return in each page of results. If there are fewer results than the value that you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1017
1029
  */
1018
1030
  MaxResults?: MaxResults;
1019
1031
  }
@@ -1041,7 +1053,7 @@ declare namespace TranscribeService {
1041
1053
  */
1042
1054
  NextToken?: NextToken;
1043
1055
  /**
1044
- * The maximum number of Call Analytics jobs to return in each page of results. If there are fewer results than the value you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1056
+ * The maximum number of Call Analytics jobs to return in each page of results. If there are fewer results than the value that you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1045
1057
  */
1046
1058
  MaxResults?: MaxResults;
1047
1059
  }
@@ -1073,7 +1085,7 @@ declare namespace TranscribeService {
1073
1085
  */
1074
1086
  NextToken?: NextToken;
1075
1087
  /**
1076
- * The maximum number of custom language models to return in each page of results. If there are fewer results than the value you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1088
+ * The maximum number of custom language models to return in each page of results. If there are fewer results than the value that you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1077
1089
  */
1078
1090
  MaxResults?: MaxResults;
1079
1091
  }
@@ -1101,7 +1113,7 @@ declare namespace TranscribeService {
1101
1113
  */
1102
1114
  NextToken?: NextToken;
1103
1115
  /**
1104
- * The maximum number of medical transcription jobs to return in each page of results. If there are fewer results than the value you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1116
+ * The maximum number of medical transcription jobs to return in each page of results. If there are fewer results than the value that you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1105
1117
  */
1106
1118
  MaxResults?: MaxResults;
1107
1119
  }
@@ -1125,11 +1137,11 @@ declare namespace TranscribeService {
1125
1137
  */
1126
1138
  NextToken?: NextToken;
1127
1139
  /**
1128
- * The maximum number of custom medical vocabularies to return in each page of results. If there are fewer results than the value you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1140
+ * The maximum number of custom medical vocabularies to return in each page of results. If there are fewer results than the value that you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1129
1141
  */
1130
1142
  MaxResults?: MaxResults;
1131
1143
  /**
1132
- * Returns only custom medical vocabularies with the specified state. Vocabularies are ordered by creation date, with the newest vocabulary first. If you don't include StateEquals, all custom medical vocabularies are returned.
1144
+ * Returns only custom medical vocabularies with the specified state. Custom vocabularies are ordered by creation date, with the newest vocabulary first. If you don't include StateEquals, all custom medical vocabularies are returned.
1133
1145
  */
1134
1146
  StateEquals?: VocabularyState;
1135
1147
  /**
@@ -1139,7 +1151,7 @@ declare namespace TranscribeService {
1139
1151
  }
1140
1152
  export interface ListMedicalVocabulariesResponse {
1141
1153
  /**
1142
- * Lists all custom medical vocabularies that have the status specified in your request. Vocabularies are ordered by creation date, with the newest vocabulary first.
1154
+ * Lists all custom medical vocabularies that have the status specified in your request. Custom vocabularies are ordered by creation date, with the newest vocabulary first.
1143
1155
  */
1144
1156
  Status?: VocabularyState;
1145
1157
  /**
@@ -1153,7 +1165,7 @@ declare namespace TranscribeService {
1153
1165
  }
1154
1166
  export interface ListTagsForResourceRequest {
1155
1167
  /**
1156
- * Returns a list of all tags associated with the specified Amazon Resource Name (ARN). ARNs have the format arn:partition:service:region:account-id:resource-type/resource-id. For example, arn:aws:transcribe:us-west-2:account-id:transcription-job/transcription-job-name. Valid values for resource-type are: transcription-job, medical-transcription-job, vocabulary, medical-vocabulary, vocabulary-filter, and language-model.
1168
+ * Returns a list of all tags associated with the specified Amazon Resource Name (ARN). ARNs have the format arn:partition:service:region:account-id:resource-type/resource-id. For example, arn:aws:transcribe:us-west-2:111122223333:transcription-job/transcription-job-name. Valid values for resource-type are: transcription-job, medical-transcription-job, vocabulary, medical-vocabulary, vocabulary-filter, and language-model.
1157
1169
  */
1158
1170
  ResourceArn: TranscribeArn;
1159
1171
  }
@@ -1181,7 +1193,7 @@ declare namespace TranscribeService {
1181
1193
  */
1182
1194
  NextToken?: NextToken;
1183
1195
  /**
1184
- * The maximum number of transcription jobs to return in each page of results. If there are fewer results than the value you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1196
+ * The maximum number of transcription jobs to return in each page of results. If there are fewer results than the value that you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1185
1197
  */
1186
1198
  MaxResults?: MaxResults;
1187
1199
  }
@@ -1205,7 +1217,7 @@ declare namespace TranscribeService {
1205
1217
  */
1206
1218
  NextToken?: NextToken;
1207
1219
  /**
1208
- * The maximum number of custom vocabularies to return in each page of results. If there are fewer results than the value you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1220
+ * The maximum number of custom vocabularies to return in each page of results. If there are fewer results than the value that you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1209
1221
  */
1210
1222
  MaxResults?: MaxResults;
1211
1223
  /**
@@ -1237,7 +1249,7 @@ declare namespace TranscribeService {
1237
1249
  */
1238
1250
  NextToken?: NextToken;
1239
1251
  /**
1240
- * The maximum number of custom vocabulary filters to return in each page of results. If there are fewer results than the value you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1252
+ * The maximum number of custom vocabulary filters to return in each page of results. If there are fewer results than the value that you specify, only the actual results are returned. If you don't specify a value, a default of 5 is used.
1241
1253
  */
1242
1254
  MaxResults?: MaxResults;
1243
1255
  /**
@@ -1264,7 +1276,7 @@ declare namespace TranscribeService {
1264
1276
  */
1265
1277
  MediaFileUri?: Uri;
1266
1278
  /**
1267
- * The Amazon S3 location of the media file you want to redact. For example: s3://DOC-EXAMPLE-BUCKET/my-media-file.flac s3://DOC-EXAMPLE-BUCKET/media-files/my-media-file.flac Note that the Amazon S3 bucket that contains your input media must be located in the same Amazon Web Services Region where you're making your transcription request. RedactedMediaFileUri is only supported for Call Analytics (StartCallAnalyticsJob) transcription requests.
1279
+ * The Amazon S3 location of the media file you want to redact. For example: s3://DOC-EXAMPLE-BUCKET/my-media-file.flac s3://DOC-EXAMPLE-BUCKET/media-files/my-media-file.flac Note that the Amazon S3 bucket that contains your input media must be located in the same Amazon Web Services Region where you're making your transcription request. RedactedMediaFileUri produces a redacted audio file in addition to a redacted transcript. It is only supported for Call Analytics (StartCallAnalyticsJob) transcription requests.
1268
1280
  */
1269
1281
  RedactedMediaFileUri?: Uri;
1270
1282
  }
@@ -1274,7 +1286,7 @@ declare namespace TranscribeService {
1274
1286
  export type MedicalMediaSampleRateHertz = number;
1275
1287
  export interface MedicalTranscript {
1276
1288
  /**
1277
- * The Amazon S3 location of your transcript. You can use this URI to access or download your transcript. If you included OutputBucketName in your transcription job request, this is the URI of that bucket. If you also included OutputKey in your request, your output is located in the path you specified in your request. If you didn't include OutputBucketName in your transcription job request, your transcript is stored in a service-managed bucket, and TranscriptFileUri provides you with a temporary URI you can use for secure access to your transcript. Temporary URIs for service-managed Amazon S3 buckets are only valid for 15 minutes. If you get an AccesDenied error, you can get a new temporary URI by running a GetTranscriptionJob or ListTranscriptionJob request.
1289
+ * The Amazon S3 location of your transcript. You can use this URI to access or download your transcript. Note that this is the Amazon S3 location you specified in your request using the OutputBucketName parameter.
1278
1290
  */
1279
1291
  TranscriptFileUri?: Uri;
1280
1292
  }
@@ -1292,7 +1304,7 @@ declare namespace TranscribeService {
1292
1304
  */
1293
1305
  LanguageCode?: LanguageCode;
1294
1306
  /**
1295
- * The sample rate, in Hertz, of the audio track in your input media file.
1307
+ * The sample rate, in hertz, of the audio track in your input media file.
1296
1308
  */
1297
1309
  MediaSampleRateHertz?: MedicalMediaSampleRateHertz;
1298
1310
  /**
@@ -1317,15 +1329,15 @@ declare namespace TranscribeService {
1317
1329
  */
1318
1330
  CompletionTime?: DateTime;
1319
1331
  /**
1320
- * If TranscriptionJobStatus is FAILED, FailureReason contains information about why the transcription job request failed. The FailureReason field contains one of the following values: Unsupported media format. The media format specified in MediaFormat isn't valid. Refer to MediaFormat for a list of supported formats. The media format provided does not match the detected media format. The media format specified in MediaFormat doesn't match the format of the input file. Check the media format of your media file and correct the specified value. Invalid sample rate for audio file. The sample rate specified in MediaSampleRateHertz isn't valid. The sample rate must be between 16,000 and 48,000 Hertz. The sample rate provided does not match the detected sample rate. The sample rate specified in MediaSampleRateHertz doesn't match the sample rate detected in your input media file. Check the sample rate of your media file and correct the specified value. Invalid file size: file size too large. The size of your media file is larger than what Amazon Transcribe can process. For more information, refer to Guidelines and quotas. Invalid number of channels: number of channels too large. Your audio contains more channels than Amazon Transcribe is able to process. For more information, refer to Guidelines and quotas.
1332
+ * If TranscriptionJobStatus is FAILED, FailureReason contains information about why the transcription job request failed. The FailureReason field contains one of the following values: Unsupported media format. The media format specified in MediaFormat isn't valid. Refer to MediaFormat for a list of supported formats. The media format provided does not match the detected media format. The media format specified in MediaFormat doesn't match the format of the input file. Check the media format of your media file and correct the specified value. Invalid sample rate for audio file. The sample rate specified in MediaSampleRateHertz isn't valid. The sample rate must be between 16,000 and 48,000 hertz. The sample rate provided does not match the detected sample rate. The sample rate specified in MediaSampleRateHertz doesn't match the sample rate detected in your input media file. Check the sample rate of your media file and correct the specified value. Invalid file size: file size too large. The size of your media file is larger than what Amazon Transcribe can process. For more information, refer to Guidelines and quotas. Invalid number of channels: number of channels too large. Your audio contains more channels than Amazon Transcribe is able to process. For more information, refer to Guidelines and quotas.
1321
1333
  */
1322
1334
  FailureReason?: FailureReason;
1323
1335
  /**
1324
- * Specify additional optional settings in your request, including channel identification, alternative transcriptions, and speaker labeling; allows you to apply custom vocabularies to your medical transcription job.
1336
+ * Provides information on any additional settings that were included in your request. Additional settings include channel identification, alternative transcriptions, speaker partitioning, custom vocabularies, and custom vocabulary filters.
1325
1337
  */
1326
1338
  Settings?: MedicalTranscriptionSetting;
1327
1339
  /**
1328
- * Labels all personal health information (PHI) identified in your transcript. For more information, see Identifying personal health information (PHI) in a transcription.
1340
+ * Indicates whether content identification was enabled for your transcription request.
1329
1341
  */
1330
1342
  ContentIdentificationType?: MedicalContentIdentificationType;
1331
1343
  /**
@@ -1390,11 +1402,11 @@ declare namespace TranscribeService {
1390
1402
  }
1391
1403
  export interface MedicalTranscriptionSetting {
1392
1404
  /**
1393
- * Enables speaker identification (diarization) in your transcription output. Speaker identification labels the speech from individual speakers in your media file. If you enable ShowSpeakerLabels in your request, you must also include MaxSpeakerLabels. You can't include both ShowSpeakerLabels and ChannelIdentification in the same request. Including both parameters returns a BadRequestException. For more information, see Identifying speakers (diarization).
1405
+ * Enables speaker partitioning (diarization) in your transcription output. Speaker partitioning labels the speech from individual speakers in your media file. If you enable ShowSpeakerLabels in your request, you must also include MaxSpeakerLabels. You can't include ShowSpeakerLabels and ChannelIdentification in the same request. Including both parameters returns a BadRequestException. For more information, see Partitioning speakers (diarization).
1394
1406
  */
1395
1407
  ShowSpeakerLabels?: Boolean;
1396
1408
  /**
1397
- * Specify the maximum number of speakers you want to identify in your media. Note that if your media contains more speakers than the specified number, multiple speakers will be identified as a single speaker. If you specify the MaxSpeakerLabels field, you must set the ShowSpeakerLabels field to true.
1409
+ * Specify the maximum number of speakers you want to partition in your media. Note that if your media contains more speakers than the specified number, multiple speakers are treated as a single speaker. If you specify the MaxSpeakerLabels field, you must set the ShowSpeakerLabels field to true.
1398
1410
  */
1399
1411
  MaxSpeakerLabels?: MaxSpeakers;
1400
1412
  /**
@@ -1410,14 +1422,14 @@ declare namespace TranscribeService {
1410
1422
  */
1411
1423
  MaxAlternatives?: MaxAlternatives;
1412
1424
  /**
1413
- * The name of the custom vocabulary you want to use when processing your medical transcription job. Vocabulary names are case sensitive. The language of the specified vocabulary must match the language code you specify in your transcription request. If the languages don't match, the vocabulary isn't applied. There are no errors or warnings associated with a language mismatch. US English (en-US) is the only valid language for Amazon Transcribe Medical.
1425
+ * The name of the custom vocabulary you want to use when processing your medical transcription job. Custom vocabulary names are case sensitive. The language of the specified custom vocabulary must match the language code that you specify in your transcription request. If the languages don't match, the custom vocabulary isn't applied. There are no errors or warnings associated with a language mismatch. US English (en-US) is the only valid language for Amazon Transcribe Medical.
1414
1426
  */
1415
1427
  VocabularyName?: VocabularyName;
1416
1428
  }
1417
1429
  export type ModelName = string;
1418
1430
  export interface ModelSettings {
1419
1431
  /**
1420
- * The name of the custom language model you want to use when processing your transcription job. Note that language model names are case sensitive. The language of the specified language model must match the language code you specify in your transcription request. If the languages don't match, the language model isn't applied. There are no errors or warnings associated with a language mismatch.
1432
+ * The name of the custom language model you want to use when processing your transcription job. Note that custom language model names are case sensitive. The language of the specified custom language model must match the language code that you specify in your transcription request. If the languages don't match, the custom language model isn't applied. There are no errors or warnings associated with a language mismatch.
1421
1433
  */
1422
1434
  LanguageModelName?: ModelName;
1423
1435
  }
@@ -1427,15 +1439,15 @@ declare namespace TranscribeService {
1427
1439
  export type NonEmptyString = string;
1428
1440
  export interface NonTalkTimeFilter {
1429
1441
  /**
1430
- * Specify the duration, in milliseconds, of the period of silence you want to flag. For example, you can flag a silent period that lasts 30000 milliseconds.
1442
+ * Specify the duration, in milliseconds, of the period of silence that you want to flag. For example, you can flag a silent period that lasts 30,000 milliseconds.
1431
1443
  */
1432
1444
  Threshold?: TimestampMilliseconds;
1433
1445
  /**
1434
- * Allows you to specify a time range (in milliseconds) in your audio, during which you want to search for a period of silence. See for more detail.
1446
+ * Makes it possible to specify a time range (in milliseconds) in your audio, during which you want to search for a period of silence. See for more detail.
1435
1447
  */
1436
1448
  AbsoluteTimeRange?: AbsoluteTimeRange;
1437
1449
  /**
1438
- * Allows you to specify a time range (in percentage) in your media file, during which you want to search for a period of silence. See for more detail.
1450
+ * Makes it possible to specify a time range (in percentage) in your media file, during which you want to search for a period of silence. See for more detail.
1439
1451
  */
1440
1452
  RelativeTimeRange?: RelativeTimeRange;
1441
1453
  /**
@@ -1464,11 +1476,11 @@ declare namespace TranscribeService {
1464
1476
  */
1465
1477
  EndPercentage?: Percentage;
1466
1478
  /**
1467
- * The time, in percentage, from the start of your media file until the value you specify in which Amazon Transcribe searches for your specified criteria.
1479
+ * The time, in percentage, from the start of your media file until the specified value. Amazon Transcribe searches for your specified criteria in this time segment.
1468
1480
  */
1469
1481
  First?: Percentage;
1470
1482
  /**
1471
- * The time, in percentage, from the value you specify until the end of your media file in which Amazon Transcribe searches for your specified criteria.
1483
+ * The time, in percentage, from the specified value until the end of your media file. Amazon Transcribe searches for your specified criteria in this time segment.
1472
1484
  */
1473
1485
  Last?: Percentage;
1474
1486
  }
@@ -1493,23 +1505,23 @@ declare namespace TranscribeService {
1493
1505
  export type RuleList = Rule[];
1494
1506
  export interface SentimentFilter {
1495
1507
  /**
1496
- * Specify the sentiments you want to flag.
1508
+ * Specify the sentiments that you want to flag.
1497
1509
  */
1498
1510
  Sentiments: SentimentValueList;
1499
1511
  /**
1500
- * Allows you to specify a time range (in milliseconds) in your audio, during which you want to search for the specified sentiments. See for more detail.
1512
+ * Makes it possible to specify a time range (in milliseconds) in your audio, during which you want to search for the specified sentiments. See for more detail.
1501
1513
  */
1502
1514
  AbsoluteTimeRange?: AbsoluteTimeRange;
1503
1515
  /**
1504
- * Allows you to specify a time range (in percentage) in your media file, during which you want to search for the specified sentiments. See for more detail.
1516
+ * Makes it possible to specify a time range (in percentage) in your media file, during which you want to search for the specified sentiments. See for more detail.
1505
1517
  */
1506
1518
  RelativeTimeRange?: RelativeTimeRange;
1507
1519
  /**
1508
- * Specify the participant you want to flag. Omitting this parameter is equivalent to specifying both participants.
1520
+ * Specify the participant that you want to flag. Omitting this parameter is equivalent to specifying both participants.
1509
1521
  */
1510
1522
  ParticipantRole?: ParticipantRole;
1511
1523
  /**
1512
- * Set to TRUE to flag the sentiments you didn't include in your request. Set to FALSE to flag the sentiments you specified in your request.
1524
+ * Set to TRUE to flag the sentiments that you didn't include in your request. Set to FALSE to flag the sentiments that you specified in your request.
1513
1525
  */
1514
1526
  Negate?: Boolean;
1515
1527
  }
@@ -1521,11 +1533,11 @@ declare namespace TranscribeService {
1521
1533
  */
1522
1534
  VocabularyName?: VocabularyName;
1523
1535
  /**
1524
- * Enables speaker identification (diarization) in your transcription output. Speaker identification labels the speech from individual speakers in your media file. If you enable ShowSpeakerLabels in your request, you must also include MaxSpeakerLabels. You can't include both ShowSpeakerLabels and ChannelIdentification in the same request. Including both parameters returns a BadRequestException. For more information, see Identifying speakers (diarization).
1536
+ * Enables speaker partitioning (diarization) in your transcription output. Speaker partitioning labels the speech from individual speakers in your media file. If you enable ShowSpeakerLabels in your request, you must also include MaxSpeakerLabels. You can't include both ShowSpeakerLabels and ChannelIdentification in the same request. Including both parameters returns a BadRequestException. For more information, see Partitioning speakers (diarization).
1525
1537
  */
1526
1538
  ShowSpeakerLabels?: Boolean;
1527
1539
  /**
1528
- * Specify the maximum number of speakers you want to identify in your media. Note that if your media contains more speakers than the specified number, multiple speakers will be identified as a single speaker. If you specify the MaxSpeakerLabels field, you must set the ShowSpeakerLabels field to true.
1540
+ * Specify the maximum number of speakers you want to partition in your media. Note that if your media contains more speakers than the specified number, multiple speakers are treated as a single speaker. If you specify the MaxSpeakerLabels field, you must set the ShowSpeakerLabels field to true.
1529
1541
  */
1530
1542
  MaxSpeakerLabels?: MaxSpeakers;
1531
1543
  /**
@@ -1545,7 +1557,7 @@ declare namespace TranscribeService {
1545
1557
  */
1546
1558
  VocabularyFilterName?: VocabularyFilterName;
1547
1559
  /**
1548
- * Specify how you want your vocabulary filter applied to your transcript. To replace words with ***, choose mask. To delete words, choose remove. To flag words without changing them, choose tag.
1560
+ * Specify how you want your custom vocabulary filter applied to your transcript. To replace words with ***, choose mask. To delete words, choose remove. To flag words without changing them, choose tag.
1549
1561
  */
1550
1562
  VocabularyFilterMethod?: VocabularyFilterMethod;
1551
1563
  }
@@ -1555,6 +1567,9 @@ declare namespace TranscribeService {
1555
1567
  * A unique name, chosen by you, for your Call Analytics job. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new job with the same name as an existing job, you get a ConflictException error.
1556
1568
  */
1557
1569
  CallAnalyticsJobName: CallAnalyticsJobName;
1570
+ /**
1571
+ * Describes the Amazon S3 location of the media file you want to use in your Call Analytics request.
1572
+ */
1558
1573
  Media: Media;
1559
1574
  /**
1560
1575
  * The Amazon S3 location where you want your Call Analytics transcription output stored. You can use any of the following formats to specify the output location: s3://DOC-EXAMPLE-BUCKET s3://DOC-EXAMPLE-BUCKET/my-output-folder/ s3://DOC-EXAMPLE-BUCKET/my-output-folder/my-call-analytics-job.json Unless you specify a file name (option 3), the name of your output file has a default value that matches the name you specified for your transcription job using the CallAnalyticsJobName parameter. You can specify a KMS key to encrypt your output using the OutputEncryptionKMSKeyId parameter. If you don't specify a KMS key, Amazon Transcribe uses the default Amazon S3 key for server-side encryption. If you don't specify OutputLocation, your transcript is placed in a service-managed Amazon S3 bucket and you are provided with a URI to access your transcript.
@@ -1565,7 +1580,7 @@ declare namespace TranscribeService {
1565
1580
  */
1566
1581
  OutputEncryptionKMSKeyId?: KMSKeyId;
1567
1582
  /**
1568
- * The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. If the role you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails. IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path. For example: arn:aws:iam::111122223333:role/Admin. For more information, see IAM ARNs.
1583
+ * The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. If the role that you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails. IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path. For example: arn:aws:iam::111122223333:role/Admin. For more information, see IAM ARNs.
1569
1584
  */
1570
1585
  DataAccessRoleArn?: DataAccessRoleArn;
1571
1586
  /**
@@ -1573,7 +1588,7 @@ declare namespace TranscribeService {
1573
1588
  */
1574
1589
  Settings?: CallAnalyticsJobSettings;
1575
1590
  /**
1576
- * Allows you to specify which speaker is on which channel. For example, if your agent is the first participant to speak, you would set ChannelId to 0 (to indicate the first channel) and ParticipantRole to AGENT (to indicate that it's the agent speaking).
1591
+ * Makes it possible to specify which speaker is on which channel. For example, if your agent is the first participant to speak, you would set ChannelId to 0 (to indicate the first channel) and ParticipantRole to AGENT (to indicate that it's the agent speaking).
1577
1592
  */
1578
1593
  ChannelDefinitions?: ChannelDefinitions;
1579
1594
  }
@@ -1585,7 +1600,7 @@ declare namespace TranscribeService {
1585
1600
  }
1586
1601
  export interface StartMedicalTranscriptionJobRequest {
1587
1602
  /**
1588
- * A unique name, chosen by you, for your medical transcription job. The name you specify is also used as the default name of your transcription output file. If you want to specify a different name for your transcription output, use the OutputKey parameter. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new job with the same name as an existing job, you get a ConflictException error.
1603
+ * A unique name, chosen by you, for your medical transcription job. The name that you specify is also used as the default name of your transcription output file. If you want to specify a different name for your transcription output, use the OutputKey parameter. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new job with the same name as an existing job, you get a ConflictException error.
1589
1604
  */
1590
1605
  MedicalTranscriptionJobName: TranscriptionJobName;
1591
1606
  /**
@@ -1593,7 +1608,7 @@ declare namespace TranscribeService {
1593
1608
  */
1594
1609
  LanguageCode: LanguageCode;
1595
1610
  /**
1596
- * The sample rate, in Hertz, of the audio track in your input media file. If you don't specify the media sample rate, Amazon Transcribe Medical determines it for you. If you specify the sample rate, it must match the rate detected by Amazon Transcribe Medical; if there's a mismatch between the value you specify and the value detected, your job fails. Therefore, in most cases, it's advised to omit MediaSampleRateHertz and let Amazon Transcribe Medical determine the sample rate.
1611
+ * The sample rate, in hertz, of the audio track in your input media file. If you don't specify the media sample rate, Amazon Transcribe Medical determines it for you. If you specify the sample rate, it must match the rate detected by Amazon Transcribe Medical; if there's a mismatch between the value that you specify and the value detected, your job fails. Therefore, in most cases, it's advised to omit MediaSampleRateHertz and let Amazon Transcribe Medical determine the sample rate.
1597
1612
  */
1598
1613
  MediaSampleRateHertz?: MedicalMediaSampleRateHertz;
1599
1614
  /**
@@ -1602,7 +1617,7 @@ declare namespace TranscribeService {
1602
1617
  MediaFormat?: MediaFormat;
1603
1618
  Media: Media;
1604
1619
  /**
1605
- * The name of the Amazon S3 bucket where you want your medical transcription output stored. Do not include the S3:// prefix of the specified bucket. If you want your output to go to a sub-folder of this bucket, specify it using the OutputKey parameter; OutputBucketName only accepts the name of a bucket. For example, if you want your output stored in S3://DOC-EXAMPLE-BUCKET, set OutputBucketName to DOC-EXAMPLE-BUCKET. However, if you want your output stored in S3://DOC-EXAMPLE-BUCKET/test-files/, set OutputBucketName to DOC-EXAMPLE-BUCKET and OutputKey to test-files/. Note that Amazon Transcribe must have permission to use the specified location. You can change Amazon S3 permissions using the Amazon Web Services Management Console. See also Permissions Required for IAM User Roles. If you don't specify OutputBucketName, your transcript is placed in a service-managed Amazon S3 bucket and you are provided with a URI to access your transcript.
1620
+ * The name of the Amazon S3 bucket where you want your medical transcription output stored. Do not include the S3:// prefix of the specified bucket. If you want your output to go to a sub-folder of this bucket, specify it using the OutputKey parameter; OutputBucketName only accepts the name of a bucket. For example, if you want your output stored in S3://DOC-EXAMPLE-BUCKET, set OutputBucketName to DOC-EXAMPLE-BUCKET. However, if you want your output stored in S3://DOC-EXAMPLE-BUCKET/test-files/, set OutputBucketName to DOC-EXAMPLE-BUCKET and OutputKey to test-files/. Note that Amazon Transcribe must have permission to use the specified location. You can change Amazon S3 permissions using the Amazon Web Services Management Console. See also Permissions Required for IAM User Roles.
1606
1621
  */
1607
1622
  OutputBucketName: OutputBucketName;
1608
1623
  /**
@@ -1618,7 +1633,7 @@ declare namespace TranscribeService {
1618
1633
  */
1619
1634
  KMSEncryptionContext?: KMSEncryptionContextMap;
1620
1635
  /**
1621
- * Specify additional optional settings in your request, including channel identification, alternative transcriptions, and speaker labeling; allows you to apply custom vocabularies to your transcription job.
1636
+ * Specify additional optional settings in your request, including channel identification, alternative transcriptions, and speaker partitioning. You can use that to apply custom vocabularies to your transcription job.
1622
1637
  */
1623
1638
  Settings?: MedicalTranscriptionSetting;
1624
1639
  /**
@@ -1646,7 +1661,7 @@ declare namespace TranscribeService {
1646
1661
  }
1647
1662
  export interface StartTranscriptionJobRequest {
1648
1663
  /**
1649
- * A unique name, chosen by you, for your transcription job. The name you specify is also used as the default name of your transcription output file. If you want to specify a different name for your transcription output, use the OutputKey parameter. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new job with the same name as an existing job, you get a ConflictException error.
1664
+ * A unique name, chosen by you, for your transcription job. The name that you specify is also used as the default name of your transcription output file. If you want to specify a different name for your transcription output, use the OutputKey parameter. This name is case sensitive, cannot contain spaces, and must be unique within an Amazon Web Services account. If you try to create a new job with the same name as an existing job, you get a ConflictException error.
1650
1665
  */
1651
1666
  TranscriptionJobName: TranscriptionJobName;
1652
1667
  /**
@@ -1654,7 +1669,7 @@ declare namespace TranscribeService {
1654
1669
  */
1655
1670
  LanguageCode?: LanguageCode;
1656
1671
  /**
1657
- * The sample rate, in Hertz, of the audio track in your input media file. If you don't specify the media sample rate, Amazon Transcribe determines it for you. If you specify the sample rate, it must match the rate detected by Amazon Transcribe; if there's a mismatch between the value you specify and the value detected, your job fails. Therefore, in most cases, it's advised to omit MediaSampleRateHertz and let Amazon Transcribe determine the sample rate.
1672
+ * The sample rate, in hertz, of the audio track in your input media file. If you don't specify the media sample rate, Amazon Transcribe determines it for you. If you specify the sample rate, it must match the rate detected by Amazon Transcribe. If there's a mismatch between the value that you specify and the value detected, your job fails. In most cases, you can omit MediaSampleRateHertz and let Amazon Transcribe determine the sample rate.
1658
1673
  */
1659
1674
  MediaSampleRateHertz?: MediaSampleRateHertz;
1660
1675
  /**
@@ -1682,7 +1697,7 @@ declare namespace TranscribeService {
1682
1697
  */
1683
1698
  KMSEncryptionContext?: KMSEncryptionContextMap;
1684
1699
  /**
1685
- * Specify additional optional settings in your request, including channel identification, alternative transcriptions, speaker labeling; allows you to apply custom vocabularies and vocabulary filters. If you want to include a custom vocabulary or a custom vocabulary filter (or both) with your request but do not want to use automatic language identification, use Settings with the VocabularyName or VocabularyFilterName (or both) sub-parameter. If you're using automatic language identification with your request and want to include a custom language model, a custom vocabulary, or a custom vocabulary filter, use instead the parameter with the LanguageModelName, VocabularyName or VocabularyFilterName sub-parameters.
1700
+ * Specify additional optional settings in your request, including channel identification, alternative transcriptions, speaker partitioning. You can use that to apply custom vocabularies and vocabulary filters. If you want to include a custom vocabulary or a custom vocabulary filter (or both) with your request but do not want to use automatic language identification, use Settings with the VocabularyName or VocabularyFilterName (or both) sub-parameter. If you're using automatic language identification with your request and want to include a custom language model, a custom vocabulary, or a custom vocabulary filter, use instead the parameter with the LanguageModelName, VocabularyName or VocabularyFilterName sub-parameters.
1686
1701
  */
1687
1702
  Settings?: Settings;
1688
1703
  /**
@@ -1690,23 +1705,23 @@ declare namespace TranscribeService {
1690
1705
  */
1691
1706
  ModelSettings?: ModelSettings;
1692
1707
  /**
1693
- * Allows you to control how your transcription job is processed. Currently, the only JobExecutionSettings modification you can choose is enabling job queueing using the AllowDeferredExecution sub-parameter. If you include JobExecutionSettings in your request, you must also include the sub-parameters: AllowDeferredExecution and DataAccessRoleArn.
1708
+ * Makes it possible to control how your transcription job is processed. Currently, the only JobExecutionSettings modification you can choose is enabling job queueing using the AllowDeferredExecution sub-parameter. If you include JobExecutionSettings in your request, you must also include the sub-parameters: AllowDeferredExecution and DataAccessRoleArn.
1694
1709
  */
1695
1710
  JobExecutionSettings?: JobExecutionSettings;
1696
1711
  /**
1697
- * Allows you to redact or flag specified personally identifiable information (PII) in your transcript. If you use ContentRedaction, you must also include the sub-parameters: PiiEntityTypes, RedactionOutput, and RedactionType.
1712
+ * Makes it possible to redact or flag specified personally identifiable information (PII) in your transcript. If you use ContentRedaction, you must also include the sub-parameters: PiiEntityTypes, RedactionOutput, and RedactionType.
1698
1713
  */
1699
1714
  ContentRedaction?: ContentRedaction;
1700
1715
  /**
1701
- * Enables automatic language identification in your transcription job request. If you include IdentifyLanguage, you can optionally include a list of language codes, using LanguageOptions, that you think may be present in your media file. Including language options can improve transcription accuracy. If you want to apply a custom language model, a custom vocabulary, or a custom vocabulary filter to your automatic language identification request, include LanguageIdSettings with the relevant sub-parameters (VocabularyName, LanguageModelName, and VocabularyFilterName). Note that you must include one of LanguageCode, IdentifyLanguage, or IdentifyMultipleLanguages in your request. If you include more than one of these parameters, your transcription job fails.
1716
+ * Enables automatic language identification in your transcription job request. Use this parameter if your media file contains only one language. If your media contains multiple languages, use IdentifyMultipleLanguages instead. If you include IdentifyLanguage, you can optionally include a list of language codes, using LanguageOptions, that you think may be present in your media file. Including LanguageOptions restricts IdentifyLanguage to only the language options that you specify, which can improve transcription accuracy. If you want to apply a custom language model, a custom vocabulary, or a custom vocabulary filter to your automatic language identification request, include LanguageIdSettings with the relevant sub-parameters (VocabularyName, LanguageModelName, and VocabularyFilterName). If you include LanguageIdSettings, also include LanguageOptions. Note that you must include one of LanguageCode, IdentifyLanguage, or IdentifyMultipleLanguages in your request. If you include more than one of these parameters, your transcription job fails.
1702
1717
  */
1703
1718
  IdentifyLanguage?: Boolean;
1704
1719
  /**
1705
- * Enables automatic multi-language identification in your transcription job request. Use this parameter if your media file contains more than one language. If you include IdentifyMultipleLanguages, you can optionally include a list of language codes, using LanguageOptions, that you think may be present in your media file. Including language options can improve transcription accuracy. If you want to apply a custom vocabulary or a custom vocabulary filter to your automatic language identification request, include LanguageIdSettings with the relevant sub-parameters (VocabularyName and VocabularyFilterName). Note that you must include one of LanguageCode, IdentifyLanguage, or IdentifyMultipleLanguages in your request. If you include more than one of these parameters, your transcription job fails.
1720
+ * Enables automatic multi-language identification in your transcription job request. Use this parameter if your media file contains more than one language. If your media contains only one language, use IdentifyLanguage instead. If you include IdentifyMultipleLanguages, you can optionally include a list of language codes, using LanguageOptions, that you think may be present in your media file. Including LanguageOptions restricts IdentifyLanguage to only the language options that you specify, which can improve transcription accuracy. If you want to apply a custom vocabulary or a custom vocabulary filter to your automatic language identification request, include LanguageIdSettings with the relevant sub-parameters (VocabularyName and VocabularyFilterName). If you include LanguageIdSettings, also include LanguageOptions. Note that you must include one of LanguageCode, IdentifyLanguage, or IdentifyMultipleLanguages in your request. If you include more than one of these parameters, your transcription job fails.
1706
1721
  */
1707
1722
  IdentifyMultipleLanguages?: Boolean;
1708
1723
  /**
1709
- * You can specify two or more language codes that represent the languages you think may be present in your media; including more than five is not recommended. If you're unsure what languages are present, do not include this parameter. If you include LanguageOptions in your request, you must also include IdentifyLanguage. For more information, refer to Supported languages. To transcribe speech in Modern Standard Arabic (ar-SA), your media file must be encoded at a sample rate of 16,000 Hz or higher.
1724
+ * You can specify two or more language codes that represent the languages you think may be present in your media. Including more than five is not recommended. If you're unsure what languages are present, do not include this parameter. If you include LanguageOptions in your request, you must also include IdentifyLanguage. For more information, refer to Supported languages. To transcribe speech in Modern Standard Arabic (ar-SA), your media file must be encoded at a sample rate of 16,000 Hz or higher.
1710
1725
  */
1711
1726
  LanguageOptions?: LanguageOptions;
1712
1727
  /**
@@ -1718,7 +1733,7 @@ declare namespace TranscribeService {
1718
1733
  */
1719
1734
  Tags?: TagList;
1720
1735
  /**
1721
- * If using automatic language identification (IdentifyLanguage) in your request and you want to apply a custom language model, a custom vocabulary, or a custom vocabulary filter, include LanguageIdSettings with the relevant sub-parameters (VocabularyName, LanguageModelName, and VocabularyFilterName). You can specify two or more language codes that represent the languages you think may be present in your media; including more than five is not recommended. Each language code you include can have an associated custom language model, custom vocabulary, and custom vocabulary filter. The languages you specify must match the languages of the specified custom language models, custom vocabularies, and custom vocabulary filters. To include language options using IdentifyLanguage without including a custom language model, a custom vocabulary, or a custom vocabulary filter, use LanguageOptions instead of LanguageIdSettings. Including language options can improve the accuracy of automatic language identification. If you want to include a custom language model with your request but do not want to use automatic language identification, use instead the parameter with the LanguageModelName sub-parameter. If you want to include a custom vocabulary or a custom vocabulary filter (or both) with your request but do not want to use automatic language identification, use instead the parameter with the VocabularyName or VocabularyFilterName (or both) sub-parameter.
1736
+ * If using automatic language identification in your request and you want to apply a custom language model, a custom vocabulary, or a custom vocabulary filter, include LanguageIdSettings with the relevant sub-parameters (VocabularyName, LanguageModelName, and VocabularyFilterName). Note that multi-language identification (IdentifyMultipleLanguages) doesn't support custom language models. LanguageIdSettings supports two to five language codes. Each language code you include can have an associated custom language model, custom vocabulary, and custom vocabulary filter. The language codes that you specify must match the languages of the associated custom language models, custom vocabularies, and custom vocabulary filters. It's recommended that you include LanguageOptions when using LanguageIdSettings to ensure that the correct language dialect is identified. For example, if you specify a custom vocabulary that is in en-US but Amazon Transcribe determines that the language spoken in your media is en-AU, your custom vocabulary is not applied to your transcription. If you include LanguageOptions and include en-US as the only English language dialect, your custom vocabulary is applied to your transcription. If you want to include a custom language model with your request but do not want to use automatic language identification, use instead the parameter with the LanguageModelName sub-parameter. If you want to include a custom vocabulary or a custom vocabulary filter (or both) with your request but do not want to use automatic language identification, use instead the parameter with the VocabularyName or VocabularyFilterName (or both) sub-parameter.
1722
1737
  */
1723
1738
  LanguageIdSettings?: LanguageIdSettingsMap;
1724
1739
  }
@@ -1772,7 +1787,7 @@ declare namespace TranscribeService {
1772
1787
  export type TagList = Tag[];
1773
1788
  export interface TagResourceRequest {
1774
1789
  /**
1775
- * The Amazon Resource Name (ARN) of the resource you want to tag. ARNs have the format arn:partition:service:region:account-id:resource-type/resource-id. For example, arn:aws:transcribe:us-west-2:account-id:transcription-job/transcription-job-name. Valid values for resource-type are: transcription-job, medical-transcription-job, vocabulary, medical-vocabulary, vocabulary-filter, and language-model.
1790
+ * The Amazon Resource Name (ARN) of the resource you want to tag. ARNs have the format arn:partition:service:region:account-id:resource-type/resource-id. For example, arn:aws:transcribe:us-west-2:111122223333:transcription-job/transcription-job-name. Valid values for resource-type are: transcription-job, medical-transcription-job, vocabulary, medical-vocabulary, vocabulary-filter, and language-model.
1776
1791
  */
1777
1792
  ResourceArn: TranscribeArn;
1778
1793
  /**
@@ -1797,27 +1812,27 @@ declare namespace TranscribeService {
1797
1812
  }
1798
1813
  export interface TranscriptFilter {
1799
1814
  /**
1800
- * Flag the presence or absence of an exact match to the phrases you specify. For example, if you specify the phrase "speak to a manager" as your Targets value, only that exact phrase is flagged. Note that semantic matching is not supported. For example, if your customer says "speak to the manager", instead of "speak to a manager", your content is not flagged.
1815
+ * Flag the presence or absence of an exact match to the phrases that you specify. For example, if you specify the phrase "speak to a manager" as your Targets value, only that exact phrase is flagged. Note that semantic matching is not supported. For example, if your customer says "speak to the manager", instead of "speak to a manager", your content is not flagged.
1801
1816
  */
1802
1817
  TranscriptFilterType: TranscriptFilterType;
1803
1818
  /**
1804
- * Allows you to specify a time range (in milliseconds) in your audio, during which you want to search for the specified key words or phrases. See for more detail.
1819
+ * Makes it possible to specify a time range (in milliseconds) in your audio, during which you want to search for the specified key words or phrases. See for more detail.
1805
1820
  */
1806
1821
  AbsoluteTimeRange?: AbsoluteTimeRange;
1807
1822
  /**
1808
- * Allows you to specify a time range (in percentage) in your media file, during which you want to search for the specified key words or phrases. See for more detail.
1823
+ * Makes it possible to specify a time range (in percentage) in your media file, during which you want to search for the specified key words or phrases. See for more detail.
1809
1824
  */
1810
1825
  RelativeTimeRange?: RelativeTimeRange;
1811
1826
  /**
1812
- * Specify the participant you want to flag. Omitting this parameter is equivalent to specifying both participants.
1827
+ * Specify the participant that you want to flag. Omitting this parameter is equivalent to specifying both participants.
1813
1828
  */
1814
1829
  ParticipantRole?: ParticipantRole;
1815
1830
  /**
1816
- * Set to TRUE to flag the absence of the phrase you specified in your request. Set to FALSE to flag the presence of the phrase you specified in your request.
1831
+ * Set to TRUE to flag the absence of the phrase that you specified in your request. Set to FALSE to flag the presence of the phrase that you specified in your request.
1817
1832
  */
1818
1833
  Negate?: Boolean;
1819
1834
  /**
1820
- * Specify the phrases you want to flag.
1835
+ * Specify the phrases that you want to flag.
1821
1836
  */
1822
1837
  Targets: StringTargetList;
1823
1838
  }
@@ -1832,11 +1847,11 @@ declare namespace TranscribeService {
1832
1847
  */
1833
1848
  TranscriptionJobStatus?: TranscriptionJobStatus;
1834
1849
  /**
1835
- * The language code used to create your transcription job. For a list of supported languages and their associated language codes, refer to the Supported languages table. Note that you must include one of LanguageCode, IdentifyLanguage, or IdentifyMultipleLanguages in your request. If you include more than one of these parameters, your transcription job fails.
1850
+ * The language code used to create your transcription job. This parameter is used with single-language identification. For multi-language identification requests, refer to the plural version of this parameter, LanguageCodes.
1836
1851
  */
1837
1852
  LanguageCode?: LanguageCode;
1838
1853
  /**
1839
- * The sample rate, in Hertz, of the audio track in your input media file.
1854
+ * The sample rate, in hertz, of the audio track in your input media file.
1840
1855
  */
1841
1856
  MediaSampleRateHertz?: MediaSampleRateHertz;
1842
1857
  /**
@@ -1844,7 +1859,7 @@ declare namespace TranscribeService {
1844
1859
  */
1845
1860
  MediaFormat?: MediaFormat;
1846
1861
  /**
1847
- * Describes the Amazon S3 location of the media file you want to use in your request.
1862
+ * Provides the Amazon S3 location of the media file you used in your request.
1848
1863
  */
1849
1864
  Media?: Media;
1850
1865
  /**
@@ -1864,23 +1879,23 @@ declare namespace TranscribeService {
1864
1879
  */
1865
1880
  CompletionTime?: DateTime;
1866
1881
  /**
1867
- * If TranscriptionJobStatus is FAILED, FailureReason contains information about why the transcription job request failed. The FailureReason field contains one of the following values: Unsupported media format. The media format specified in MediaFormat isn't valid. Refer to MediaFormat for a list of supported formats. The media format provided does not match the detected media format. The media format specified in MediaFormat doesn't match the format of the input file. Check the media format of your media file and correct the specified value. Invalid sample rate for audio file. The sample rate specified in MediaSampleRateHertz isn't valid. The sample rate must be between 8,000 and 48,000 Hertz. The sample rate provided does not match the detected sample rate. The sample rate specified in MediaSampleRateHertz doesn't match the sample rate detected in your input media file. Check the sample rate of your media file and correct the specified value. Invalid file size: file size too large. The size of your media file is larger than what Amazon Transcribe can process. For more information, refer to Guidelines and quotas. Invalid number of channels: number of channels too large. Your audio contains more channels than Amazon Transcribe is able to process. For more information, refer to Guidelines and quotas.
1882
+ * If TranscriptionJobStatus is FAILED, FailureReason contains information about why the transcription job request failed. The FailureReason field contains one of the following values: Unsupported media format. The media format specified in MediaFormat isn't valid. Refer to MediaFormat for a list of supported formats. The media format provided does not match the detected media format. The media format specified in MediaFormat doesn't match the format of the input file. Check the media format of your media file and correct the specified value. Invalid sample rate for audio file. The sample rate specified in MediaSampleRateHertz isn't valid. The sample rate must be between 8,000 and 48,000 hertz. The sample rate provided does not match the detected sample rate. The sample rate specified in MediaSampleRateHertz doesn't match the sample rate detected in your input media file. Check the sample rate of your media file and correct the specified value. Invalid file size: file size too large. The size of your media file is larger than what Amazon Transcribe can process. For more information, refer to Guidelines and quotas. Invalid number of channels: number of channels too large. Your audio contains more channels than Amazon Transcribe is able to process. For more information, refer to Guidelines and quotas.
1868
1883
  */
1869
1884
  FailureReason?: FailureReason;
1870
1885
  /**
1871
- * Specify additional optional settings in your request, including channel identification, alternative transcriptions, speaker labeling; allows you to apply custom vocabularies and vocabulary filters. If you want to include a custom vocabulary or a custom vocabulary filter (or both) with your request but do not want to use automatic language identification, use Settings with the VocabularyName or VocabularyFilterName (or both) sub-parameter. If you're using automatic language identification with your request and want to include a custom language model, a custom vocabulary, or a custom vocabulary filter, do not use the Settings parameter; use instead the parameter with the LanguageModelName, VocabularyName or VocabularyFilterName sub-parameters.
1886
+ * Provides information on any additional settings that were included in your request. Additional settings include channel identification, alternative transcriptions, speaker partitioning, custom vocabularies, and custom vocabulary filters.
1872
1887
  */
1873
1888
  Settings?: Settings;
1874
1889
  /**
1875
- * The custom language model you want to include with your transcription job. If you include ModelSettings in your request, you must include the LanguageModelName sub-parameter.
1890
+ * Provides information on the custom language model you included in your request.
1876
1891
  */
1877
1892
  ModelSettings?: ModelSettings;
1878
1893
  /**
1879
- * Provides information about how your transcription job is being processed. This parameter shows if your request is queued and what data access role is being used.
1894
+ * Provides information about how your transcription job was processed. This parameter shows if your request was queued and what data access role was used.
1880
1895
  */
1881
1896
  JobExecutionSettings?: JobExecutionSettings;
1882
1897
  /**
1883
- * Redacts or flags specified personally identifiable information (PII) in your transcript.
1898
+ * Indicates whether redaction was enabled in your transcript.
1884
1899
  */
1885
1900
  ContentRedaction?: ContentRedaction;
1886
1901
  /**
@@ -1892,7 +1907,7 @@ declare namespace TranscribeService {
1892
1907
  */
1893
1908
  IdentifyMultipleLanguages?: Boolean;
1894
1909
  /**
1895
- * You can specify two or more language codes that represent the languages you think may be present in your media; including more than five is not recommended. If you're unsure what languages are present, do not include this parameter. If you include LanguageOptions in your request, you must also include IdentifyLanguage. For more information, refer to Supported languages. To transcribe speech in Modern Standard Arabic (ar-SA), your media file must be encoded at a sample rate of 16,000 Hz or higher.
1910
+ * Provides the language codes you specified in your request.
1896
1911
  */
1897
1912
  LanguageOptions?: LanguageOptions;
1898
1913
  /**
@@ -1900,19 +1915,19 @@ declare namespace TranscribeService {
1900
1915
  */
1901
1916
  IdentifiedLanguageScore?: IdentifiedLanguageScore;
1902
1917
  /**
1903
- * The language codes used to create your transcription job. This parameter is used with multi-language identification. For single-language identification requests, refer to the singular version of this parameter, LanguageCode. For a list of supported languages and their associated language codes, refer to the Supported languages table.
1918
+ * The language codes used to create your transcription job. This parameter is used with multi-language identification. For single-language identification requests, refer to the singular version of this parameter, LanguageCode.
1904
1919
  */
1905
1920
  LanguageCodes?: LanguageCodeList;
1906
1921
  /**
1907
- * Adds one or more custom tags, each in the form of a key:value pair, to a new transcription job at the time you start this new job. To learn more about using tags with Amazon Transcribe, refer to Tagging resources.
1922
+ * The tags, each in the form of a key:value pair, assigned to the specified transcription job.
1908
1923
  */
1909
1924
  Tags?: TagList;
1910
1925
  /**
1911
- * Generate subtitles for your media file with your transcription request.
1926
+ * Indicates whether subtitles were generated with your transcription.
1912
1927
  */
1913
1928
  Subtitles?: SubtitlesOutput;
1914
1929
  /**
1915
- * If using automatic language identification (IdentifyLanguage) in your request and you want to apply a custom language model, a custom vocabulary, or a custom vocabulary filter, include LanguageIdSettings with the relevant sub-parameters (VocabularyName, LanguageModelName, and VocabularyFilterName). You can specify two or more language codes that represent the languages you think may be present in your media; including more than five is not recommended. Each language code you include can have an associated custom language model, custom vocabulary, and custom vocabulary filter. The languages you specify must match the languages of the specified custom language models, custom vocabularies, and custom vocabulary filters. To include language options using IdentifyLanguage without including a custom language model, a custom vocabulary, or a custom vocabulary filter, use LanguageOptions instead of LanguageIdSettings. Including language options can improve the accuracy of automatic language identification. If you want to include a custom language model with your request but do not want to use automatic language identification, use instead the parameter with the LanguageModelName sub-parameter. If you want to include a custom vocabulary or a custom vocabulary filter (or both) with your request but do not want to use automatic language identification, use instead the parameter with the VocabularyName or VocabularyFilterName (or both) sub-parameter.
1930
+ * Provides the name and language of all custom language models, custom vocabularies, and custom vocabulary filters that you included in your request.
1916
1931
  */
1917
1932
  LanguageIdSettings?: LanguageIdSettingsMap;
1918
1933
  }
@@ -1977,7 +1992,7 @@ declare namespace TranscribeService {
1977
1992
  export type Type = "CONVERSATION"|"DICTATION"|string;
1978
1993
  export interface UntagResourceRequest {
1979
1994
  /**
1980
- * The Amazon Resource Name (ARN) of the Amazon Transcribe resource you want to remove tags from. ARNs have the format arn:partition:service:region:account-id:resource-type/resource-id. For example, arn:aws:transcribe:us-west-2:account-id:transcription-job/transcription-job-name. Valid values for resource-type are: transcription-job, medical-transcription-job, vocabulary, medical-vocabulary, vocabulary-filter, and language-model.
1995
+ * The Amazon Resource Name (ARN) of the Amazon Transcribe resource you want to remove tags from. ARNs have the format arn:partition:service:region:account-id:resource-type/resource-id. For example, arn:aws:transcribe:us-west-2:111122223333:transcription-job/transcription-job-name. Valid values for resource-type are: transcription-job, medical-transcription-job, vocabulary, medical-vocabulary, vocabulary-filter, and language-model.
1981
1996
  */
1982
1997
  ResourceArn: TranscribeArn;
1983
1998
  /**
@@ -1996,6 +2011,10 @@ declare namespace TranscribeService {
1996
2011
  * The rules used for the updated Call Analytics category. The rules you provide in this field replace the ones that are currently being used in the specified category.
1997
2012
  */
1998
2013
  Rules: RuleList;
2014
+ /**
2015
+ * Choose whether you want to update a streaming or a batch Call Analytics category. The input type you specify must match the input type specified when the category was created. For example, if you created a category with the POST_CALL input type, you must use POST_CALL as the input type when updating this category.
2016
+ */
2017
+ InputType?: InputType;
1999
2018
  }
2000
2019
  export interface UpdateCallAnalyticsCategoryResponse {
2001
2020
  /**
@@ -2005,7 +2024,7 @@ declare namespace TranscribeService {
2005
2024
  }
2006
2025
  export interface UpdateMedicalVocabularyRequest {
2007
2026
  /**
2008
- * The name of the custom medical vocabulary you want to update. Vocabulary names are case sensitive.
2027
+ * The name of the custom medical vocabulary you want to update. Custom medical vocabulary names are case sensitive.
2009
2028
  */
2010
2029
  VocabularyName: VocabularyName;
2011
2030
  /**
@@ -2023,7 +2042,7 @@ declare namespace TranscribeService {
2023
2042
  */
2024
2043
  VocabularyName?: VocabularyName;
2025
2044
  /**
2026
- * The language code you selected for your medical vocabulary. US English (en-US) is the only language supported with Amazon Transcribe Medical.
2045
+ * The language code you selected for your custom medical vocabulary. US English (en-US) is the only language supported with Amazon Transcribe Medical.
2027
2046
  */
2028
2047
  LanguageCode?: LanguageCode;
2029
2048
  /**
@@ -2031,17 +2050,17 @@ declare namespace TranscribeService {
2031
2050
  */
2032
2051
  LastModifiedTime?: DateTime;
2033
2052
  /**
2034
- * The processing state of your custom medical vocabulary. If the state is READY, you can use the vocabulary in a StartMedicalTranscriptionJob request.
2053
+ * The processing state of your custom medical vocabulary. If the state is READY, you can use the custom vocabulary in a StartMedicalTranscriptionJob request.
2035
2054
  */
2036
2055
  VocabularyState?: VocabularyState;
2037
2056
  }
2038
2057
  export interface UpdateVocabularyFilterRequest {
2039
2058
  /**
2040
- * The name of the custom vocabulary filter you want to update. Vocabulary filter names are case sensitive.
2059
+ * The name of the custom vocabulary filter you want to update. Custom vocabulary filter names are case sensitive.
2041
2060
  */
2042
2061
  VocabularyFilterName: VocabularyFilterName;
2043
2062
  /**
2044
- * Use this parameter if you want to update your vocabulary filter by including all desired terms, as comma-separated values, within your request. The other option for updating your vocabulary filter is to save your entries in a text file and upload them to an Amazon S3 bucket, then specify the location of your file using the VocabularyFilterFileUri parameter. Note that if you include Words in your request, you cannot use VocabularyFilterFileUri; you must choose one or the other. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language.
2063
+ * Use this parameter if you want to update your custom vocabulary filter by including all desired terms, as comma-separated values, within your request. The other option for updating your vocabulary filter is to save your entries in a text file and upload them to an Amazon S3 bucket, then specify the location of your file using the VocabularyFilterFileUri parameter. Note that if you include Words in your request, you cannot use VocabularyFilterFileUri; you must choose one or the other. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language.
2045
2064
  */
2046
2065
  Words?: Words;
2047
2066
  /**
@@ -2055,25 +2074,25 @@ declare namespace TranscribeService {
2055
2074
  */
2056
2075
  VocabularyFilterName?: VocabularyFilterName;
2057
2076
  /**
2058
- * The language code you selected for your vocabulary filter.
2077
+ * The language code you selected for your custom vocabulary filter.
2059
2078
  */
2060
2079
  LanguageCode?: LanguageCode;
2061
2080
  /**
2062
- * The date and time the specified vocabulary filter was last updated. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
2081
+ * The date and time the specified custom vocabulary filter was last updated. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
2063
2082
  */
2064
2083
  LastModifiedTime?: DateTime;
2065
2084
  }
2066
2085
  export interface UpdateVocabularyRequest {
2067
2086
  /**
2068
- * The name of the custom vocabulary you want to update. Vocabulary names are case sensitive.
2087
+ * The name of the custom vocabulary you want to update. Custom vocabulary names are case sensitive.
2069
2088
  */
2070
2089
  VocabularyName: VocabularyName;
2071
2090
  /**
2072
- * The language code that represents the language of the entries in the custom vocabulary you want to update. Each vocabulary must contain terms in only one language. A custom vocabulary can only be used to transcribe files in the same language as the vocabulary. For example, if you create a vocabulary using US English (en-US), you can only apply this vocabulary to files that contain English audio. For a list of supported languages and their associated language codes, refer to the Supported languages table.
2091
+ * The language code that represents the language of the entries in the custom vocabulary you want to update. Each custom vocabulary must contain terms in only one language. A custom vocabulary can only be used to transcribe files in the same language as the custom vocabulary. For example, if you create a custom vocabulary using US English (en-US), you can only apply this custom vocabulary to files that contain English audio. For a list of supported languages and their associated language codes, refer to the Supported languages table.
2073
2092
  */
2074
2093
  LanguageCode: LanguageCode;
2075
2094
  /**
2076
- * Use this parameter if you want to update your vocabulary by including all desired terms, as comma-separated values, within your request. The other option for updating your vocabulary is to save your entries in a text file and upload them to an Amazon S3 bucket, then specify the location of your file using the VocabularyFileUri parameter. Note that if you include Phrases in your request, you cannot use VocabularyFileUri; you must choose one or the other. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language.
2095
+ * Use this parameter if you want to update your custom vocabulary by including all desired terms, as comma-separated values, within your request. The other option for updating your custom vocabulary is to save your entries in a text file and upload them to an Amazon S3 bucket, then specify the location of your file using the VocabularyFileUri parameter. Note that if you include Phrases in your request, you cannot use VocabularyFileUri; you must choose one or the other. Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language.
2077
2096
  */
2078
2097
  Phrases?: Phrases;
2079
2098
  /**
@@ -2091,11 +2110,11 @@ declare namespace TranscribeService {
2091
2110
  */
2092
2111
  LanguageCode?: LanguageCode;
2093
2112
  /**
2094
- * The date and time the specified vocabulary was last updated. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
2113
+ * The date and time the specified custom vocabulary was last updated. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
2095
2114
  */
2096
2115
  LastModifiedTime?: DateTime;
2097
2116
  /**
2098
- * The processing state of your custom vocabulary. If the state is READY, you can use the vocabulary in a StartTranscriptionJob request.
2117
+ * The processing state of your custom vocabulary. If the state is READY, you can use the custom vocabulary in a StartTranscriptionJob request.
2099
2118
  */
2100
2119
  VocabularyState?: VocabularyState;
2101
2120
  }
@@ -2107,11 +2126,11 @@ declare namespace TranscribeService {
2107
2126
  */
2108
2127
  VocabularyFilterName?: VocabularyFilterName;
2109
2128
  /**
2110
- * The language code that represents the language of the entries in your vocabulary filter. Each vocabulary filter must contain terms in only one language. A vocabulary filter can only be used to transcribe files in the same language as the filter. For example, if you create a vocabulary filter using US English (en-US), you can only apply this filter to files that contain English audio. For a list of supported languages and their associated language codes, refer to the Supported languages table.
2129
+ * The language code that represents the language of the entries in your vocabulary filter. Each custom vocabulary filter must contain terms in only one language. A custom vocabulary filter can only be used to transcribe files in the same language as the filter. For example, if you create a custom vocabulary filter using US English (en-US), you can only apply this filter to files that contain English audio. For a list of supported languages and their associated language codes, refer to the Supported languages table.
2111
2130
  */
2112
2131
  LanguageCode?: LanguageCode;
2113
2132
  /**
2114
- * The date and time the specified vocabulary filter was last modified. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
2133
+ * The date and time the specified custom vocabulary filter was last modified. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
2115
2134
  */
2116
2135
  LastModifiedTime?: DateTime;
2117
2136
  }
@@ -2124,15 +2143,15 @@ declare namespace TranscribeService {
2124
2143
  */
2125
2144
  VocabularyName?: VocabularyName;
2126
2145
  /**
2127
- * The language code used to create your custom vocabulary. Each vocabulary must contain terms in only one language. A custom vocabulary can only be used to transcribe files in the same language as the vocabulary. For example, if you create a vocabulary using US English (en-US), you can only apply this vocabulary to files that contain English audio.
2146
+ * The language code used to create your custom vocabulary. Each custom vocabulary must contain terms in only one language. A custom vocabulary can only be used to transcribe files in the same language as the custom vocabulary. For example, if you create a custom vocabulary using US English (en-US), you can only apply this custom vocabulary to files that contain English audio.
2128
2147
  */
2129
2148
  LanguageCode?: LanguageCode;
2130
2149
  /**
2131
- * The date and time the specified vocabulary was last modified. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
2150
+ * The date and time the specified custom vocabulary was last modified. Timestamps are in the format YYYY-MM-DD'T'HH:MM:SS.SSSSSS-UTC. For example, 2022-05-04T12:32:58.761000-07:00 represents 12:32 PM UTC-7 on May 4, 2022.
2132
2151
  */
2133
2152
  LastModifiedTime?: DateTime;
2134
2153
  /**
2135
- * The processing state of your custom vocabulary. If the state is READY, you can use the vocabulary in a StartTranscriptionJob request.
2154
+ * The processing state of your custom vocabulary. If the state is READY, you can use the custom vocabulary in a StartTranscriptionJob request.
2136
2155
  */
2137
2156
  VocabularyState?: VocabularyState;
2138
2157
  }