aws-sdk-transcribestreamingservice 1.44.0 → 1.46.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +12 -0
- data/VERSION +1 -1
- data/lib/aws-sdk-transcribestreamingservice/async_client.rb +457 -12
- data/lib/aws-sdk-transcribestreamingservice/client.rb +1 -1
- data/lib/aws-sdk-transcribestreamingservice/client_api.rb +165 -0
- data/lib/aws-sdk-transcribestreamingservice/endpoint_provider.rb +47 -116
- data/lib/aws-sdk-transcribestreamingservice/endpoints.rb +14 -0
- data/lib/aws-sdk-transcribestreamingservice/event_streams.rb +124 -0
- data/lib/aws-sdk-transcribestreamingservice/plugins/endpoints.rb +2 -0
- data/lib/aws-sdk-transcribestreamingservice/types.rb +769 -71
- data/lib/aws-sdk-transcribestreamingservice.rb +2 -2
- metadata +2 -2
@@ -47,13 +47,6 @@ module Aws::TranscribeStreamingService
|
|
47
47
|
#
|
48
48
|
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/event-stream.html
|
49
49
|
#
|
50
|
-
# @note When making an API call, you may pass AudioEvent
|
51
|
-
# data as a hash:
|
52
|
-
#
|
53
|
-
# {
|
54
|
-
# audio_chunk: "data",
|
55
|
-
# }
|
56
|
-
#
|
57
50
|
# @!attribute [rw] audio_chunk
|
58
51
|
# An audio blob that contains the next part of the audio that you want
|
59
52
|
# to transcribe. The maximum audio chunk size is 32 KB.
|
@@ -68,8 +61,9 @@ module Aws::TranscribeStreamingService
|
|
68
61
|
include Aws::Structure
|
69
62
|
end
|
70
63
|
|
71
|
-
# One or more arguments to the `StartStreamTranscription
|
72
|
-
# `StartMedicalStreamTranscription
|
64
|
+
# One or more arguments to the `StartStreamTranscription`,
|
65
|
+
# `StartMedicalStreamTranscription`, or
|
66
|
+
# `StartCallAnalyticsStreamTranscription` operation was not valid. For
|
73
67
|
# example, `MediaEncoding` or `LanguageCode` used not valid values.
|
74
68
|
# Check the specified parameters and try your request again.
|
75
69
|
#
|
@@ -85,6 +79,209 @@ module Aws::TranscribeStreamingService
|
|
85
79
|
include Aws::Structure
|
86
80
|
end
|
87
81
|
|
82
|
+
# Contains entities identified as personally identifiable information
|
83
|
+
# (PII) in your transcription output, along with various associated
|
84
|
+
# attributes. Examples include category, confidence score, content,
|
85
|
+
# type, and start and end times.
|
86
|
+
#
|
87
|
+
# @!attribute [rw] begin_offset_millis
|
88
|
+
# The time, in milliseconds, from the beginning of the audio stream to
|
89
|
+
# the start of the identified entity.
|
90
|
+
# @return [Integer]
|
91
|
+
#
|
92
|
+
# @!attribute [rw] end_offset_millis
|
93
|
+
# The time, in milliseconds, from the beginning of the audio stream to
|
94
|
+
# the end of the identified entity.
|
95
|
+
# @return [Integer]
|
96
|
+
#
|
97
|
+
# @!attribute [rw] category
|
98
|
+
# The category of information identified. For example, `PII`.
|
99
|
+
# @return [String]
|
100
|
+
#
|
101
|
+
# @!attribute [rw] type
|
102
|
+
# The type of PII identified. For example, `NAME` or
|
103
|
+
# `CREDIT_DEBIT_NUMBER`.
|
104
|
+
# @return [String]
|
105
|
+
#
|
106
|
+
# @!attribute [rw] content
|
107
|
+
# The word or words that represent the identified entity.
|
108
|
+
# @return [String]
|
109
|
+
#
|
110
|
+
# @!attribute [rw] confidence
|
111
|
+
# The confidence score associated with the identification of an entity
|
112
|
+
# in your transcript.
|
113
|
+
#
|
114
|
+
# Confidence scores are values between 0 and 1. A larger value
|
115
|
+
# indicates a higher probability that the identified entity correctly
|
116
|
+
# matches the entity spoken in your media.
|
117
|
+
# @return [Float]
|
118
|
+
#
|
119
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/CallAnalyticsEntity AWS API Documentation
|
120
|
+
#
|
121
|
+
class CallAnalyticsEntity < Struct.new(
|
122
|
+
:begin_offset_millis,
|
123
|
+
:end_offset_millis,
|
124
|
+
:category,
|
125
|
+
:type,
|
126
|
+
:content,
|
127
|
+
:confidence)
|
128
|
+
SENSITIVE = []
|
129
|
+
include Aws::Structure
|
130
|
+
end
|
131
|
+
|
132
|
+
# A word, phrase, or punctuation mark in your Call Analytics
|
133
|
+
# transcription output, along with various associated attributes, such
|
134
|
+
# as confidence score, type, and start and end times.
|
135
|
+
#
|
136
|
+
# @!attribute [rw] begin_offset_millis
|
137
|
+
# The time, in milliseconds, from the beginning of the audio stream to
|
138
|
+
# the start of the identified item.
|
139
|
+
# @return [Integer]
|
140
|
+
#
|
141
|
+
# @!attribute [rw] end_offset_millis
|
142
|
+
# The time, in milliseconds, from the beginning of the audio stream to
|
143
|
+
# the end of the identified item.
|
144
|
+
# @return [Integer]
|
145
|
+
#
|
146
|
+
# @!attribute [rw] type
|
147
|
+
# The type of item identified. Options are: `PRONUNCIATION` (spoken
|
148
|
+
# words) and `PUNCTUATION`.
|
149
|
+
# @return [String]
|
150
|
+
#
|
151
|
+
# @!attribute [rw] content
|
152
|
+
# The word or punctuation that was transcribed.
|
153
|
+
# @return [String]
|
154
|
+
#
|
155
|
+
# @!attribute [rw] confidence
|
156
|
+
# The confidence score associated with a word or phrase in your
|
157
|
+
# transcript.
|
158
|
+
#
|
159
|
+
# Confidence scores are values between 0 and 1. A larger value
|
160
|
+
# indicates a higher probability that the identified item correctly
|
161
|
+
# matches the item spoken in your media.
|
162
|
+
# @return [Float]
|
163
|
+
#
|
164
|
+
# @!attribute [rw] vocabulary_filter_match
|
165
|
+
# Indicates whether the specified item matches a word in the
|
166
|
+
# vocabulary filter included in your Call Analytics request. If
|
167
|
+
# `true`, there is a vocabulary filter match.
|
168
|
+
# @return [Boolean]
|
169
|
+
#
|
170
|
+
# @!attribute [rw] stable
|
171
|
+
# If partial result stabilization is enabled, `Stable` indicates
|
172
|
+
# whether the specified item is stable (`true`) or if it may change
|
173
|
+
# when the segment is complete (`false`).
|
174
|
+
# @return [Boolean]
|
175
|
+
#
|
176
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/CallAnalyticsItem AWS API Documentation
|
177
|
+
#
|
178
|
+
class CallAnalyticsItem < Struct.new(
|
179
|
+
:begin_offset_millis,
|
180
|
+
:end_offset_millis,
|
181
|
+
:type,
|
182
|
+
:content,
|
183
|
+
:confidence,
|
184
|
+
:vocabulary_filter_match,
|
185
|
+
:stable)
|
186
|
+
SENSITIVE = []
|
187
|
+
include Aws::Structure
|
188
|
+
end
|
189
|
+
|
190
|
+
# Provides information on any `TranscriptFilterType` categories that
|
191
|
+
# matched your transcription output. Matches are identified for each
|
192
|
+
# segment upon completion of that segment.
|
193
|
+
#
|
194
|
+
# @!attribute [rw] matched_categories
|
195
|
+
# Lists the categories that were matched in your audio segment.
|
196
|
+
# @return [Array<String>]
|
197
|
+
#
|
198
|
+
# @!attribute [rw] matched_details
|
199
|
+
# Contains information about the matched categories, including
|
200
|
+
# category names and timestamps.
|
201
|
+
# @return [Hash<String,Types::PointsOfInterest>]
|
202
|
+
#
|
203
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/CategoryEvent AWS API Documentation
|
204
|
+
#
|
205
|
+
class CategoryEvent < Struct.new(
|
206
|
+
:matched_categories,
|
207
|
+
:matched_details,
|
208
|
+
:event_type)
|
209
|
+
SENSITIVE = []
|
210
|
+
include Aws::Structure
|
211
|
+
end
|
212
|
+
|
213
|
+
# Makes it possible to specify which speaker is on which audio channel.
|
214
|
+
# For example, if your agent is the first participant to speak, you
|
215
|
+
# would set `ChannelId` to `0` (to indicate the first channel) and
|
216
|
+
# `ParticipantRole` to `AGENT` (to indicate that it's the agent
|
217
|
+
# speaking).
|
218
|
+
#
|
219
|
+
# @!attribute [rw] channel_id
|
220
|
+
# Specify the audio channel you want to define.
|
221
|
+
# @return [Integer]
|
222
|
+
#
|
223
|
+
# @!attribute [rw] participant_role
|
224
|
+
# Specify the speaker you want to define. Omitting this parameter is
|
225
|
+
# equivalent to specifying both participants.
|
226
|
+
# @return [String]
|
227
|
+
#
|
228
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/ChannelDefinition AWS API Documentation
|
229
|
+
#
|
230
|
+
class ChannelDefinition < Struct.new(
|
231
|
+
:channel_id,
|
232
|
+
:participant_role)
|
233
|
+
SENSITIVE = []
|
234
|
+
include Aws::Structure
|
235
|
+
end
|
236
|
+
|
237
|
+
# Provides the location, using character count, in your transcript where
|
238
|
+
# a match is identified. For example, the location of an issue or a
|
239
|
+
# category match within a segment.
|
240
|
+
#
|
241
|
+
# @!attribute [rw] begin
|
242
|
+
# Provides the character count of the first character where a match is
|
243
|
+
# identified. For example, the first character associated with an
|
244
|
+
# issue or a category match in a segment transcript.
|
245
|
+
# @return [Integer]
|
246
|
+
#
|
247
|
+
# @!attribute [rw] end
|
248
|
+
# Provides the character count of the last character where a match is
|
249
|
+
# identified. For example, the last character associated with an issue
|
250
|
+
# or a category match in a segment transcript.
|
251
|
+
# @return [Integer]
|
252
|
+
#
|
253
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/CharacterOffsets AWS API Documentation
|
254
|
+
#
|
255
|
+
class CharacterOffsets < Struct.new(
|
256
|
+
:begin,
|
257
|
+
:end)
|
258
|
+
SENSITIVE = []
|
259
|
+
include Aws::Structure
|
260
|
+
end
|
261
|
+
|
262
|
+
# Allows you to set audio channel definitions and post-call analytics
|
263
|
+
# settings.
|
264
|
+
#
|
265
|
+
# @!attribute [rw] channel_definitions
|
266
|
+
# Indicates which speaker is on which audio channel.
|
267
|
+
# @return [Array<Types::ChannelDefinition>]
|
268
|
+
#
|
269
|
+
# @!attribute [rw] post_call_analytics_settings
|
270
|
+
# Provides additional optional settings for your Call Analytics
|
271
|
+
# post-call request, including encryption and output locations for
|
272
|
+
# your redacted and unredacted transcript.
|
273
|
+
# @return [Types::PostCallAnalyticsSettings]
|
274
|
+
#
|
275
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/ConfigurationEvent AWS API Documentation
|
276
|
+
#
|
277
|
+
class ConfigurationEvent < Struct.new(
|
278
|
+
:channel_definitions,
|
279
|
+
:post_call_analytics_settings,
|
280
|
+
:event_type)
|
281
|
+
SENSITIVE = []
|
282
|
+
include Aws::Structure
|
283
|
+
end
|
284
|
+
|
88
285
|
# A new stream started with the same session ID. The current stream has
|
89
286
|
# been terminated.
|
90
287
|
#
|
@@ -165,6 +362,21 @@ module Aws::TranscribeStreamingService
|
|
165
362
|
include Aws::Structure
|
166
363
|
end
|
167
364
|
|
365
|
+
# Lists the issues that were identified in your audio segment.
|
366
|
+
#
|
367
|
+
# @!attribute [rw] character_offsets
|
368
|
+
# Provides the timestamps that identify when in an audio segment the
|
369
|
+
# specified issue occurs.
|
370
|
+
# @return [Types::CharacterOffsets]
|
371
|
+
#
|
372
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/IssueDetected AWS API Documentation
|
373
|
+
#
|
374
|
+
class IssueDetected < Struct.new(
|
375
|
+
:character_offsets)
|
376
|
+
SENSITIVE = []
|
377
|
+
include Aws::Structure
|
378
|
+
end
|
379
|
+
|
168
380
|
# A word, phrase, or punctuation mark in your transcription output,
|
169
381
|
# along with various associated attributes, such as confidence score,
|
170
382
|
# type, and start and end times.
|
@@ -486,6 +698,112 @@ module Aws::TranscribeStreamingService
|
|
486
698
|
include Aws::Structure
|
487
699
|
end
|
488
700
|
|
701
|
+
# Contains the timestamps of matched categories.
|
702
|
+
#
|
703
|
+
# @!attribute [rw] timestamp_ranges
|
704
|
+
# Contains the timestamp ranges (start time through end time) of
|
705
|
+
# matched categories and rules.
|
706
|
+
# @return [Array<Types::TimestampRange>]
|
707
|
+
#
|
708
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/PointsOfInterest AWS API Documentation
|
709
|
+
#
|
710
|
+
class PointsOfInterest < Struct.new(
|
711
|
+
:timestamp_ranges)
|
712
|
+
SENSITIVE = []
|
713
|
+
include Aws::Structure
|
714
|
+
end
|
715
|
+
|
716
|
+
# Allows you to specify additional settings for your streaming Call
|
717
|
+
# Analytics post-call request, including output locations for your
|
718
|
+
# redacted and unredacted transcript, which IAM role to use, and,
|
719
|
+
# optionally, which encryption key to use.
|
720
|
+
#
|
721
|
+
# `ContentRedactionOutput`, `DataAccessRoleArn`, and `OutputLocation`
|
722
|
+
# are required fields.
|
723
|
+
#
|
724
|
+
# @!attribute [rw] output_location
|
725
|
+
# The Amazon S3 location where you want your Call Analytics post-call
|
726
|
+
# transcription output stored. You can use any of the following
|
727
|
+
# formats to specify the output location:
|
728
|
+
#
|
729
|
+
# 1. s3://DOC-EXAMPLE-BUCKET
|
730
|
+
#
|
731
|
+
# 2. s3://DOC-EXAMPLE-BUCKET/my-output-folder/
|
732
|
+
#
|
733
|
+
# 3. s3://DOC-EXAMPLE-BUCKET/my-output-folder/my-call-analytics-job.json
|
734
|
+
# @return [String]
|
735
|
+
#
|
736
|
+
# @!attribute [rw] data_access_role_arn
|
737
|
+
# The Amazon Resource Name (ARN) of an IAM role that has permissions
|
738
|
+
# to access the Amazon S3 bucket that contains your input files. If
|
739
|
+
# the role that you specify doesn’t have the appropriate permissions
|
740
|
+
# to access the specified Amazon S3 location, your request fails.
|
741
|
+
#
|
742
|
+
# IAM role ARNs have the format
|
743
|
+
# `arn:partition:iam::account:role/role-name-with-path`. For example:
|
744
|
+
# `arn:aws:iam::111122223333:role/Admin`. For more information, see
|
745
|
+
# [IAM ARNs][1].
|
746
|
+
#
|
747
|
+
#
|
748
|
+
#
|
749
|
+
# [1]: https://docs.aws.amazon.com/IAM/latest/UserGuide/reference_identifiers.html#identifiers-arns
|
750
|
+
# @return [String]
|
751
|
+
#
|
752
|
+
# @!attribute [rw] content_redaction_output
|
753
|
+
# Specify whether you want only a redacted transcript or both a
|
754
|
+
# redacted and an unredacted transcript. If you choose redacted and
|
755
|
+
# unredacted, two JSON files are generated and stored in the Amazon S3
|
756
|
+
# output location you specify.
|
757
|
+
#
|
758
|
+
# Note that to include `ContentRedactionOutput` in your request, you
|
759
|
+
# must enable content redaction (`ContentRedactionType`).
|
760
|
+
# @return [String]
|
761
|
+
#
|
762
|
+
# @!attribute [rw] output_encryption_kms_key_id
|
763
|
+
# The KMS key you want to use to encrypt your Call Analytics post-call
|
764
|
+
# output.
|
765
|
+
#
|
766
|
+
# If using a key located in the **current** Amazon Web Services
|
767
|
+
# account, you can specify your KMS key in one of four ways:
|
768
|
+
#
|
769
|
+
# 1. Use the KMS key ID itself. For example,
|
770
|
+
# `1234abcd-12ab-34cd-56ef-1234567890ab`.
|
771
|
+
#
|
772
|
+
# 2. Use an alias for the KMS key ID. For example,
|
773
|
+
# `alias/ExampleAlias`.
|
774
|
+
#
|
775
|
+
# 3. Use the Amazon Resource Name (ARN) for the KMS key ID. For
|
776
|
+
# example,
|
777
|
+
# `arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab`.
|
778
|
+
#
|
779
|
+
# 4. Use the ARN for the KMS key alias. For example,
|
780
|
+
# `arn:aws:kms:region:account-ID:alias/ExampleAlias`.
|
781
|
+
#
|
782
|
+
# If using a key located in a **different** Amazon Web Services
|
783
|
+
# account than the current Amazon Web Services account, you can
|
784
|
+
# specify your KMS key in one of two ways:
|
785
|
+
#
|
786
|
+
# 1. Use the ARN for the KMS key ID. For example,
|
787
|
+
# `arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab`.
|
788
|
+
#
|
789
|
+
# 2. Use the ARN for the KMS key alias. For example,
|
790
|
+
# `arn:aws:kms:region:account-ID:alias/ExampleAlias`.
|
791
|
+
#
|
792
|
+
# Note that the user making the request must have permission to use
|
793
|
+
# the specified KMS key.
|
794
|
+
# @return [String]
|
795
|
+
#
|
796
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/PostCallAnalyticsSettings AWS API Documentation
|
797
|
+
#
|
798
|
+
class PostCallAnalyticsSettings < Struct.new(
|
799
|
+
:output_location,
|
800
|
+
:data_access_role_arn,
|
801
|
+
:content_redaction_output,
|
802
|
+
:output_encryption_kms_key_id)
|
803
|
+
SENSITIVE = []
|
804
|
+
include Aws::Structure
|
805
|
+
end
|
806
|
+
|
489
807
|
# The `Result` associated with a ``.
|
490
808
|
#
|
491
809
|
# Contains a set of transcription results from one or more audio
|
@@ -520,7 +838,7 @@ module Aws::TranscribeStreamingService
|
|
520
838
|
# @return [Array<Types::Alternative>]
|
521
839
|
#
|
522
840
|
# @!attribute [rw] channel_id
|
523
|
-
# Indicates
|
841
|
+
# Indicates which audio channel is associated with the `Result`.
|
524
842
|
# @return [String]
|
525
843
|
#
|
526
844
|
# @!attribute [rw] language_code
|
@@ -565,24 +883,322 @@ module Aws::TranscribeStreamingService
|
|
565
883
|
include Aws::Structure
|
566
884
|
end
|
567
885
|
|
568
|
-
#
|
569
|
-
#
|
570
|
-
#
|
571
|
-
#
|
572
|
-
#
|
573
|
-
#
|
574
|
-
#
|
575
|
-
#
|
576
|
-
#
|
577
|
-
#
|
578
|
-
#
|
579
|
-
#
|
580
|
-
#
|
581
|
-
#
|
582
|
-
#
|
583
|
-
#
|
584
|
-
#
|
886
|
+
# @!attribute [rw] language_code
|
887
|
+
# Specify the language code that represents the language spoken in
|
888
|
+
# your audio.
|
889
|
+
#
|
890
|
+
# If you're unsure of the language spoken in your audio, consider
|
891
|
+
# using `IdentifyLanguage` to enable automatic language
|
892
|
+
# identification.
|
893
|
+
#
|
894
|
+
# For a list of languages supported with streaming Call Analytics,
|
895
|
+
# refer to the [Supported languages][1] table.
|
896
|
+
#
|
897
|
+
#
|
898
|
+
#
|
899
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/supported-languages.html
|
900
|
+
# @return [String]
|
901
|
+
#
|
902
|
+
# @!attribute [rw] media_sample_rate_hertz
|
903
|
+
# The sample rate of the input audio (in hertz). Low-quality audio,
|
904
|
+
# such as telephone audio, is typically around 8,000 Hz. High-quality
|
905
|
+
# audio typically ranges from 16,000 Hz to 48,000 Hz. Note that the
|
906
|
+
# sample rate you specify must match that of your audio.
|
907
|
+
# @return [Integer]
|
908
|
+
#
|
909
|
+
# @!attribute [rw] media_encoding
|
910
|
+
# Specify the encoding of your input audio. Supported formats are:
|
911
|
+
#
|
912
|
+
# * FLAC
|
913
|
+
#
|
914
|
+
# * OPUS-encoded audio in an Ogg container
|
915
|
+
#
|
916
|
+
# * PCM (only signed 16-bit little-endian audio formats, which does
|
917
|
+
# not include WAV)
|
918
|
+
#
|
919
|
+
# For more information, see [Media formats][1].
|
920
|
+
#
|
921
|
+
#
|
922
|
+
#
|
923
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/how-input.html#how-input-audio
|
924
|
+
# @return [String]
|
925
|
+
#
|
926
|
+
# @!attribute [rw] vocabulary_name
|
927
|
+
# Specify the name of the custom vocabulary that you want to use when
|
928
|
+
# processing your transcription. Note that vocabulary names are case
|
929
|
+
# sensitive.
|
930
|
+
#
|
931
|
+
# If the language of the specified custom vocabulary doesn't match
|
932
|
+
# the language identified in your media, the custom vocabulary is not
|
933
|
+
# applied to your transcription.
|
934
|
+
#
|
935
|
+
# For more information, see [Custom vocabularies][1].
|
936
|
+
#
|
937
|
+
#
|
938
|
+
#
|
939
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/custom-vocabulary.html
|
940
|
+
# @return [String]
|
941
|
+
#
|
942
|
+
# @!attribute [rw] session_id
|
943
|
+
# Specify a name for your Call Analytics transcription session. If you
|
944
|
+
# don't include this parameter in your request, Amazon Transcribe
|
945
|
+
# generates an ID and returns it in the response.
|
946
|
+
#
|
947
|
+
# You can use a session ID to retry a streaming session.
|
948
|
+
# @return [String]
|
949
|
+
#
|
950
|
+
# @!attribute [rw] audio_stream
|
951
|
+
# An encoded stream of audio blobs. Audio streams are encoded as
|
952
|
+
# either HTTP/2 or WebSocket data frames.
|
953
|
+
#
|
954
|
+
# For more information, see [Transcribing streaming audio][1].
|
955
|
+
#
|
956
|
+
#
|
957
|
+
#
|
958
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/streaming.html
|
959
|
+
# @return [Types::AudioStream]
|
960
|
+
#
|
961
|
+
# @!attribute [rw] vocabulary_filter_name
|
962
|
+
# Specify the name of the custom vocabulary filter that you want to
|
963
|
+
# use when processing your transcription. Note that vocabulary filter
|
964
|
+
# names are case sensitive.
|
965
|
+
#
|
966
|
+
# If the language of the specified custom vocabulary filter doesn't
|
967
|
+
# match the language identified in your media, the vocabulary filter
|
968
|
+
# is not applied to your transcription.
|
969
|
+
#
|
970
|
+
# For more information, see [Using vocabulary filtering with unwanted
|
971
|
+
# words][1].
|
972
|
+
#
|
973
|
+
#
|
974
|
+
#
|
975
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/vocabulary-filtering.html
|
976
|
+
# @return [String]
|
977
|
+
#
|
978
|
+
# @!attribute [rw] vocabulary_filter_method
|
979
|
+
# Specify how you want your vocabulary filter applied to your
|
980
|
+
# transcript.
|
981
|
+
#
|
982
|
+
# To replace words with `***`, choose `mask`.
|
983
|
+
#
|
984
|
+
# To delete words, choose `remove`.
|
985
|
+
#
|
986
|
+
# To flag words without changing them, choose `tag`.
|
987
|
+
# @return [String]
|
988
|
+
#
|
989
|
+
# @!attribute [rw] language_model_name
|
990
|
+
# Specify the name of the custom language model that you want to use
|
991
|
+
# when processing your transcription. Note that language model names
|
992
|
+
# are case sensitive.
|
993
|
+
#
|
994
|
+
# The language of the specified language model must match the language
|
995
|
+
# code you specify in your transcription request. If the languages
|
996
|
+
# don't match, the custom language model isn't applied. There are no
|
997
|
+
# errors or warnings associated with a language mismatch.
|
998
|
+
#
|
999
|
+
# For more information, see [Custom language models][1].
|
1000
|
+
#
|
1001
|
+
#
|
1002
|
+
#
|
1003
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/custom-language-models.html
|
1004
|
+
# @return [String]
|
1005
|
+
#
|
1006
|
+
# @!attribute [rw] enable_partial_results_stabilization
|
1007
|
+
# Enables partial result stabilization for your transcription. Partial
|
1008
|
+
# result stabilization can reduce latency in your output, but may
|
1009
|
+
# impact accuracy. For more information, see [Partial-result
|
1010
|
+
# stabilization][1].
|
1011
|
+
#
|
1012
|
+
#
|
1013
|
+
#
|
1014
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/streaming.html#streaming-partial-result-stabilization
|
1015
|
+
# @return [Boolean]
|
1016
|
+
#
|
1017
|
+
# @!attribute [rw] partial_results_stability
|
1018
|
+
# Specify the level of stability to use when you enable partial
|
1019
|
+
# results stabilization (`EnablePartialResultsStabilization`).
|
1020
|
+
#
|
1021
|
+
# Low stability provides the highest accuracy. High stability
|
1022
|
+
# transcribes faster, but with slightly lower accuracy.
|
1023
|
+
#
|
1024
|
+
# For more information, see [Partial-result stabilization][1].
|
1025
|
+
#
|
1026
|
+
#
|
1027
|
+
#
|
1028
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/streaming.html#streaming-partial-result-stabilization
|
1029
|
+
# @return [String]
|
1030
|
+
#
|
1031
|
+
# @!attribute [rw] content_identification_type
|
1032
|
+
# Labels all personally identifiable information (PII) identified in
|
1033
|
+
# your transcript.
|
1034
|
+
#
|
1035
|
+
# Content identification is performed at the segment level; PII
|
1036
|
+
# specified in `PiiEntityTypes` is flagged upon complete transcription
|
1037
|
+
# of an audio segment.
|
1038
|
+
#
|
1039
|
+
# You can’t set `ContentIdentificationType` and `ContentRedactionType`
|
1040
|
+
# in the same request. If you set both, your request returns a
|
1041
|
+
# `BadRequestException`.
|
1042
|
+
#
|
1043
|
+
# For more information, see [Redacting or identifying personally
|
1044
|
+
# identifiable information][1].
|
1045
|
+
#
|
1046
|
+
#
|
1047
|
+
#
|
1048
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/pii-redaction.html
|
1049
|
+
# @return [String]
|
1050
|
+
#
|
1051
|
+
# @!attribute [rw] content_redaction_type
|
1052
|
+
# Redacts all personally identifiable information (PII) identified in
|
1053
|
+
# your transcript.
|
1054
|
+
#
|
1055
|
+
# Content redaction is performed at the segment level; PII specified
|
1056
|
+
# in `PiiEntityTypes` is redacted upon complete transcription of an
|
1057
|
+
# audio segment.
|
1058
|
+
#
|
1059
|
+
# You can’t set `ContentRedactionType` and `ContentIdentificationType`
|
1060
|
+
# in the same request. If you set both, your request returns a
|
1061
|
+
# `BadRequestException`.
|
1062
|
+
#
|
1063
|
+
# For more information, see [Redacting or identifying personally
|
1064
|
+
# identifiable information][1].
|
1065
|
+
#
|
1066
|
+
#
|
1067
|
+
#
|
1068
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/pii-redaction.html
|
1069
|
+
# @return [String]
|
1070
|
+
#
|
1071
|
+
# @!attribute [rw] pii_entity_types
|
1072
|
+
# Specify which types of personally identifiable information (PII) you
|
1073
|
+
# want to redact in your transcript. You can include as many types as
|
1074
|
+
# you'd like, or you can select `ALL`.
|
1075
|
+
#
|
1076
|
+
# To include `PiiEntityTypes` in your Call Analytics request, you must
|
1077
|
+
# also include either `ContentIdentificationType` or
|
1078
|
+
# `ContentRedactionType`.
|
1079
|
+
#
|
1080
|
+
# Values must be comma-separated and can include:
|
1081
|
+
# `BANK_ACCOUNT_NUMBER`, `BANK_ROUTING`, `CREDIT_DEBIT_NUMBER`,
|
1082
|
+
# `CREDIT_DEBIT_CVV`, `CREDIT_DEBIT_EXPIRY`, `PIN`, `EMAIL`,
|
1083
|
+
# `ADDRESS`, `NAME`, `PHONE`, `SSN`, or `ALL`.
|
1084
|
+
# @return [String]
|
1085
|
+
#
|
1086
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/StartCallAnalyticsStreamTranscriptionRequest AWS API Documentation
|
1087
|
+
#
|
1088
|
+
class StartCallAnalyticsStreamTranscriptionRequest < Struct.new(
|
1089
|
+
:language_code,
|
1090
|
+
:media_sample_rate_hertz,
|
1091
|
+
:media_encoding,
|
1092
|
+
:vocabulary_name,
|
1093
|
+
:session_id,
|
1094
|
+
:audio_stream,
|
1095
|
+
:vocabulary_filter_name,
|
1096
|
+
:vocabulary_filter_method,
|
1097
|
+
:language_model_name,
|
1098
|
+
:enable_partial_results_stabilization,
|
1099
|
+
:partial_results_stability,
|
1100
|
+
:content_identification_type,
|
1101
|
+
:content_redaction_type,
|
1102
|
+
:pii_entity_types)
|
1103
|
+
SENSITIVE = []
|
1104
|
+
include Aws::Structure
|
1105
|
+
end
|
1106
|
+
|
1107
|
+
# @!attribute [rw] request_id
|
1108
|
+
# Provides the identifier for your Call Analytics streaming request.
|
1109
|
+
# @return [String]
|
585
1110
|
#
|
1111
|
+
# @!attribute [rw] language_code
|
1112
|
+
# Provides the language code that you specified in your Call Analytics
|
1113
|
+
# request.
|
1114
|
+
# @return [String]
|
1115
|
+
#
|
1116
|
+
# @!attribute [rw] media_sample_rate_hertz
|
1117
|
+
# Provides the sample rate that you specified in your Call Analytics
|
1118
|
+
# request.
|
1119
|
+
# @return [Integer]
|
1120
|
+
#
|
1121
|
+
# @!attribute [rw] media_encoding
|
1122
|
+
# Provides the media encoding you specified in your Call Analytics
|
1123
|
+
# request.
|
1124
|
+
# @return [String]
|
1125
|
+
#
|
1126
|
+
# @!attribute [rw] vocabulary_name
|
1127
|
+
# Provides the name of the custom vocabulary that you specified in
|
1128
|
+
# your Call Analytics request.
|
1129
|
+
# @return [String]
|
1130
|
+
#
|
1131
|
+
# @!attribute [rw] session_id
|
1132
|
+
# Provides the identifier for your Call Analytics transcription
|
1133
|
+
# session.
|
1134
|
+
# @return [String]
|
1135
|
+
#
|
1136
|
+
# @!attribute [rw] call_analytics_transcript_result_stream
|
1137
|
+
# Provides detailed information about your Call Analytics streaming
|
1138
|
+
# session.
|
1139
|
+
# @return [Types::CallAnalyticsTranscriptResultStream]
|
1140
|
+
#
|
1141
|
+
# @!attribute [rw] vocabulary_filter_name
|
1142
|
+
# Provides the name of the custom vocabulary filter that you specified
|
1143
|
+
# in your Call Analytics request.
|
1144
|
+
# @return [String]
|
1145
|
+
#
|
1146
|
+
# @!attribute [rw] vocabulary_filter_method
|
1147
|
+
# Provides the vocabulary filtering method used in your Call Analytics
|
1148
|
+
# transcription.
|
1149
|
+
# @return [String]
|
1150
|
+
#
|
1151
|
+
# @!attribute [rw] language_model_name
|
1152
|
+
# Provides the name of the custom language model that you specified in
|
1153
|
+
# your Call Analytics request.
|
1154
|
+
# @return [String]
|
1155
|
+
#
|
1156
|
+
# @!attribute [rw] enable_partial_results_stabilization
|
1157
|
+
# Shows whether partial results stabilization was enabled for your
|
1158
|
+
# Call Analytics transcription.
|
1159
|
+
# @return [Boolean]
|
1160
|
+
#
|
1161
|
+
# @!attribute [rw] partial_results_stability
|
1162
|
+
# Provides the stabilization level used for your transcription.
|
1163
|
+
# @return [String]
|
1164
|
+
#
|
1165
|
+
# @!attribute [rw] content_identification_type
|
1166
|
+
# Shows whether content identification was enabled for your Call
|
1167
|
+
# Analytics transcription.
|
1168
|
+
# @return [String]
|
1169
|
+
#
|
1170
|
+
# @!attribute [rw] content_redaction_type
|
1171
|
+
# Shows whether content redaction was enabled for your Call Analytics
|
1172
|
+
# transcription.
|
1173
|
+
# @return [String]
|
1174
|
+
#
|
1175
|
+
# @!attribute [rw] pii_entity_types
|
1176
|
+
# Lists the PII entity types you specified in your Call Analytics
|
1177
|
+
# request.
|
1178
|
+
# @return [String]
|
1179
|
+
#
|
1180
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/StartCallAnalyticsStreamTranscriptionResponse AWS API Documentation
|
1181
|
+
#
|
1182
|
+
class StartCallAnalyticsStreamTranscriptionResponse < Struct.new(
|
1183
|
+
:request_id,
|
1184
|
+
:language_code,
|
1185
|
+
:media_sample_rate_hertz,
|
1186
|
+
:media_encoding,
|
1187
|
+
:vocabulary_name,
|
1188
|
+
:session_id,
|
1189
|
+
:call_analytics_transcript_result_stream,
|
1190
|
+
:vocabulary_filter_name,
|
1191
|
+
:vocabulary_filter_method,
|
1192
|
+
:language_model_name,
|
1193
|
+
:enable_partial_results_stabilization,
|
1194
|
+
:partial_results_stability,
|
1195
|
+
:content_identification_type,
|
1196
|
+
:content_redaction_type,
|
1197
|
+
:pii_entity_types)
|
1198
|
+
SENSITIVE = []
|
1199
|
+
include Aws::Structure
|
1200
|
+
end
|
1201
|
+
|
586
1202
|
# @!attribute [rw] language_code
|
587
1203
|
# Specify the language code that represents the language spoken in
|
588
1204
|
# your audio.
|
@@ -795,34 +1411,6 @@ module Aws::TranscribeStreamingService
|
|
795
1411
|
include Aws::Structure
|
796
1412
|
end
|
797
1413
|
|
798
|
-
# @note When making an API call, you may pass StartStreamTranscriptionRequest
|
799
|
-
# data as a hash:
|
800
|
-
#
|
801
|
-
# {
|
802
|
-
# language_code: "en-US", # accepts en-US, en-GB, es-US, fr-CA, fr-FR, en-AU, it-IT, de-DE, pt-BR, ja-JP, ko-KR, zh-CN, hi-IN, th-TH
|
803
|
-
# media_sample_rate_hertz: 1, # required
|
804
|
-
# media_encoding: "pcm", # required, accepts pcm, ogg-opus, flac
|
805
|
-
# vocabulary_name: "VocabularyName",
|
806
|
-
# session_id: "SessionId",
|
807
|
-
# input_event_stream_hander: EventStreams::AudioStream.new,
|
808
|
-
# vocabulary_filter_name: "VocabularyFilterName",
|
809
|
-
# vocabulary_filter_method: "remove", # accepts remove, mask, tag
|
810
|
-
# show_speaker_label: false,
|
811
|
-
# enable_channel_identification: false,
|
812
|
-
# number_of_channels: 1,
|
813
|
-
# enable_partial_results_stabilization: false,
|
814
|
-
# partial_results_stability: "high", # accepts high, medium, low
|
815
|
-
# content_identification_type: "PII", # accepts PII
|
816
|
-
# content_redaction_type: "PII", # accepts PII
|
817
|
-
# pii_entity_types: "PiiEntityTypes",
|
818
|
-
# language_model_name: "ModelName",
|
819
|
-
# identify_language: false,
|
820
|
-
# language_options: "LanguageOptions",
|
821
|
-
# preferred_language: "en-US", # accepts en-US, en-GB, es-US, fr-CA, fr-FR, en-AU, it-IT, de-DE, pt-BR, ja-JP, ko-KR, zh-CN, hi-IN, th-TH
|
822
|
-
# vocabulary_names: "VocabularyNames",
|
823
|
-
# vocabulary_filter_names: "VocabularyFilterNames",
|
824
|
-
# }
|
825
|
-
#
|
826
1414
|
# @!attribute [rw] language_code
|
827
1415
|
# Specify the language code that represents the language spoken in
|
828
1416
|
# your audio.
|
@@ -847,8 +1435,7 @@ module Aws::TranscribeStreamingService
|
|
847
1435
|
# @return [Integer]
|
848
1436
|
#
|
849
1437
|
# @!attribute [rw] media_encoding
|
850
|
-
# Specify the encoding
|
851
|
-
# are:
|
1438
|
+
# Specify the encoding of your input audio. Supported formats are:
|
852
1439
|
#
|
853
1440
|
# * FLAC
|
854
1441
|
#
|
@@ -870,7 +1457,8 @@ module Aws::TranscribeStreamingService
|
|
870
1457
|
# sensitive.
|
871
1458
|
#
|
872
1459
|
# If the language of the specified custom vocabulary doesn't match
|
873
|
-
# the language identified in your media,
|
1460
|
+
# the language identified in your media, the custom vocabulary is not
|
1461
|
+
# applied to your transcription.
|
874
1462
|
#
|
875
1463
|
# This parameter is **not** intended for use with the
|
876
1464
|
# `IdentifyLanguage` parameter. If you're including
|
@@ -910,7 +1498,8 @@ module Aws::TranscribeStreamingService
|
|
910
1498
|
# names are case sensitive.
|
911
1499
|
#
|
912
1500
|
# If the language of the specified custom vocabulary filter doesn't
|
913
|
-
# match the language identified in your media,
|
1501
|
+
# match the language identified in your media, the vocabulary filter
|
1502
|
+
# is not applied to your transcription.
|
914
1503
|
#
|
915
1504
|
# This parameter is **not** intended for use with the
|
916
1505
|
# `IdentifyLanguage` parameter. If you're including
|
@@ -1058,8 +1647,8 @@ module Aws::TranscribeStreamingService
|
|
1058
1647
|
#
|
1059
1648
|
# The language of the specified language model must match the language
|
1060
1649
|
# code you specify in your transcription request. If the languages
|
1061
|
-
# don't match, the language model isn't applied. There are no
|
1062
|
-
# or warnings associated with a language mismatch.
|
1650
|
+
# don't match, the custom language model isn't applied. There are no
|
1651
|
+
# errors or warnings associated with a language mismatch.
|
1063
1652
|
#
|
1064
1653
|
# For more information, see [Custom language models][1].
|
1065
1654
|
#
|
@@ -1329,6 +1918,28 @@ module Aws::TranscribeStreamingService
|
|
1329
1918
|
include Aws::Structure
|
1330
1919
|
end
|
1331
1920
|
|
1921
|
+
# Contains the timestamp range (start time through end time) of a
|
1922
|
+
# matched category.
|
1923
|
+
#
|
1924
|
+
# @!attribute [rw] begin_offset_millis
|
1925
|
+
# The time, in milliseconds, from the beginning of the audio stream to
|
1926
|
+
# the start of the category match.
|
1927
|
+
# @return [Integer]
|
1928
|
+
#
|
1929
|
+
# @!attribute [rw] end_offset_millis
|
1930
|
+
# The time, in milliseconds, from the beginning of the audio stream to
|
1931
|
+
# the end of the category match.
|
1932
|
+
# @return [Integer]
|
1933
|
+
#
|
1934
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/TimestampRange AWS API Documentation
|
1935
|
+
#
|
1936
|
+
class TimestampRange < Struct.new(
|
1937
|
+
:begin_offset_millis,
|
1938
|
+
:end_offset_millis)
|
1939
|
+
SENSITIVE = []
|
1940
|
+
include Aws::Structure
|
1941
|
+
end
|
1942
|
+
|
1332
1943
|
# The `Transcript` associated with a `.</p> Transcript contains
|
1333
1944
|
# Results, which contains a set of transcription results from one or
|
1334
1945
|
# more audio segments, along with additional information per your
|
@@ -1376,6 +1987,76 @@ module Aws::TranscribeStreamingService
|
|
1376
1987
|
include Aws::Structure
|
1377
1988
|
end
|
1378
1989
|
|
1990
|
+
# Contains set of transcription results from one or more audio segments,
|
1991
|
+
# along with additional information about the parameters included in
|
1992
|
+
# your request. For example, channel definitions, partial result
|
1993
|
+
# stabilization, sentiment, and issue detection.
|
1994
|
+
#
|
1995
|
+
# @!attribute [rw] utterance_id
|
1996
|
+
# The unique identifier that is associated with the specified
|
1997
|
+
# `UtteranceEvent`.
|
1998
|
+
# @return [String]
|
1999
|
+
#
|
2000
|
+
# @!attribute [rw] is_partial
|
2001
|
+
# Indicates whether the segment in the `UtteranceEvent` is complete
|
2002
|
+
# (`FALSE`) or partial (`TRUE`).
|
2003
|
+
# @return [Boolean]
|
2004
|
+
#
|
2005
|
+
# @!attribute [rw] participant_role
|
2006
|
+
# Provides the role of the speaker for each audio channel, either
|
2007
|
+
# `CUSTOMER` or `AGENT`.
|
2008
|
+
# @return [String]
|
2009
|
+
#
|
2010
|
+
# @!attribute [rw] begin_offset_millis
|
2011
|
+
# The time, in milliseconds, from the beginning of the audio stream to
|
2012
|
+
# the start of the `UtteranceEvent`.
|
2013
|
+
# @return [Integer]
|
2014
|
+
#
|
2015
|
+
# @!attribute [rw] end_offset_millis
|
2016
|
+
# The time, in milliseconds, from the beginning of the audio stream to
|
2017
|
+
# the start of the `UtteranceEvent`.
|
2018
|
+
# @return [Integer]
|
2019
|
+
#
|
2020
|
+
# @!attribute [rw] transcript
|
2021
|
+
# Contains transcribed text.
|
2022
|
+
# @return [String]
|
2023
|
+
#
|
2024
|
+
# @!attribute [rw] items
|
2025
|
+
# Contains words, phrases, or punctuation marks that are associated
|
2026
|
+
# with the specified `UtteranceEvent`.
|
2027
|
+
# @return [Array<Types::CallAnalyticsItem>]
|
2028
|
+
#
|
2029
|
+
# @!attribute [rw] entities
|
2030
|
+
# Contains entities identified as personally identifiable information
|
2031
|
+
# (PII) in your transcription output.
|
2032
|
+
# @return [Array<Types::CallAnalyticsEntity>]
|
2033
|
+
#
|
2034
|
+
# @!attribute [rw] sentiment
|
2035
|
+
# Provides the sentiment that was detected in the specified segment.
|
2036
|
+
# @return [String]
|
2037
|
+
#
|
2038
|
+
# @!attribute [rw] issues_detected
|
2039
|
+
# Provides the issue that was detected in the specified segment.
|
2040
|
+
# @return [Array<Types::IssueDetected>]
|
2041
|
+
#
|
2042
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/UtteranceEvent AWS API Documentation
|
2043
|
+
#
|
2044
|
+
class UtteranceEvent < Struct.new(
|
2045
|
+
:utterance_id,
|
2046
|
+
:is_partial,
|
2047
|
+
:participant_role,
|
2048
|
+
:begin_offset_millis,
|
2049
|
+
:end_offset_millis,
|
2050
|
+
:transcript,
|
2051
|
+
:items,
|
2052
|
+
:entities,
|
2053
|
+
:sentiment,
|
2054
|
+
:issues_detected,
|
2055
|
+
:event_type)
|
2056
|
+
SENSITIVE = []
|
2057
|
+
include Aws::Structure
|
2058
|
+
end
|
2059
|
+
|
1379
2060
|
# An encoded stream of audio blobs. Audio streams are encoded as either
|
1380
2061
|
# HTTP/2 or WebSocket data frames.
|
1381
2062
|
#
|
@@ -1385,15 +2066,6 @@ module Aws::TranscribeStreamingService
|
|
1385
2066
|
#
|
1386
2067
|
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/streaming.html
|
1387
2068
|
#
|
1388
|
-
# @note When making an API call, you may pass AudioStream
|
1389
|
-
# data as a hash:
|
1390
|
-
#
|
1391
|
-
# {
|
1392
|
-
# audio_event: {
|
1393
|
-
# audio_chunk: "data",
|
1394
|
-
# },
|
1395
|
-
# }
|
1396
|
-
#
|
1397
2069
|
# EventStream is an Enumerator of Events.
|
1398
2070
|
# #event_types #=> Array, returns all modeled event types in the stream
|
1399
2071
|
#
|
@@ -1403,7 +2075,33 @@ module Aws::TranscribeStreamingService
|
|
1403
2075
|
|
1404
2076
|
def event_types
|
1405
2077
|
[
|
1406
|
-
:audio_event
|
2078
|
+
:audio_event,
|
2079
|
+
:configuration_event
|
2080
|
+
]
|
2081
|
+
end
|
2082
|
+
|
2083
|
+
end
|
2084
|
+
|
2085
|
+
# Contains detailed information about your Call Analytics streaming
|
2086
|
+
# session. These details are provided in the `UtteranceEvent` and
|
2087
|
+
# `CategoryEvent` objects.
|
2088
|
+
#
|
2089
|
+
# EventStream is an Enumerator of Events.
|
2090
|
+
# #event_types #=> Array, returns all modeled event types in the stream
|
2091
|
+
#
|
2092
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/CallAnalyticsTranscriptResultStream AWS API Documentation
|
2093
|
+
#
|
2094
|
+
class CallAnalyticsTranscriptResultStream < Enumerator
|
2095
|
+
|
2096
|
+
def event_types
|
2097
|
+
[
|
2098
|
+
:utterance_event,
|
2099
|
+
:category_event,
|
2100
|
+
:bad_request_exception,
|
2101
|
+
:limit_exceeded_exception,
|
2102
|
+
:internal_failure_exception,
|
2103
|
+
:conflict_exception,
|
2104
|
+
:service_unavailable_exception
|
1407
2105
|
]
|
1408
2106
|
end
|
1409
2107
|
|