aws-sdk-transcribestreamingservice 1.44.0 → 1.46.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +12 -0
- data/VERSION +1 -1
- data/lib/aws-sdk-transcribestreamingservice/async_client.rb +457 -12
- data/lib/aws-sdk-transcribestreamingservice/client.rb +1 -1
- data/lib/aws-sdk-transcribestreamingservice/client_api.rb +165 -0
- data/lib/aws-sdk-transcribestreamingservice/endpoint_provider.rb +47 -116
- data/lib/aws-sdk-transcribestreamingservice/endpoints.rb +14 -0
- data/lib/aws-sdk-transcribestreamingservice/event_streams.rb +124 -0
- data/lib/aws-sdk-transcribestreamingservice/plugins/endpoints.rb +2 -0
- data/lib/aws-sdk-transcribestreamingservice/types.rb +769 -71
- data/lib/aws-sdk-transcribestreamingservice.rb +2 -2
- metadata +2 -2
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 5ddffcae0c324d67b4b6af46536e2bef70f43e36058bd67188d441fe8c11550d
|
4
|
+
data.tar.gz: b1d33c6d9c90eaf302c16e45fbcd0aa14e70e02abe21e762cd346baccb6fcece
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 2678ed0b27e6906a094b47673bb37719469b08eeddd86184f330a9d14e01610f14b38c683aba290cd3c114ae7f756d026e54e64d4a1dc8a8805ae010ac748baa
|
7
|
+
data.tar.gz: f1e407492143b88822586890a5f840c0b7acf9b0d3250fb46eec39cd499e493ee18683f2c69998a928c3dd3f1fce9b826fa1d2e6cb419fbd16c90cbc51a6b133
|
data/CHANGELOG.md
CHANGED
@@ -1,6 +1,18 @@
|
|
1
1
|
Unreleased Changes
|
2
2
|
------------------
|
3
3
|
|
4
|
+
1.46.0 (2023-01-18)
|
5
|
+
------------------
|
6
|
+
|
7
|
+
* Feature - Code Generated Changes, see `./build_tools` or `aws-sdk-core`'s CHANGELOG.md for details.
|
8
|
+
|
9
|
+
* Issue - Replace runtime endpoint resolution approach with generated ruby code.
|
10
|
+
|
11
|
+
1.45.0 (2022-11-28)
|
12
|
+
------------------
|
13
|
+
|
14
|
+
* Feature - This release adds support for real-time (streaming) and post-call Call Analytics within Amazon Transcribe.
|
15
|
+
|
4
16
|
1.44.0 (2022-11-09)
|
5
17
|
------------------
|
6
18
|
|
data/VERSION
CHANGED
@@ -1 +1 @@
|
|
1
|
-
1.
|
1
|
+
1.46.0
|
@@ -271,10 +271,453 @@ module Aws::TranscribeStreamingService
|
|
271
271
|
|
272
272
|
# @!group API Operations
|
273
273
|
|
274
|
+
# Starts a bidirectional HTTP/2 or WebSocket stream where audio is
|
275
|
+
# streamed to Amazon Transcribe and the transcription results are
|
276
|
+
# streamed to your application. Use this operation for [Call
|
277
|
+
# Analytics][1] transcriptions.
|
278
|
+
#
|
279
|
+
# The following parameters are required:
|
280
|
+
#
|
281
|
+
# * `language-code`
|
282
|
+
#
|
283
|
+
# * `media-encoding`
|
284
|
+
#
|
285
|
+
# * `sample-rate`
|
286
|
+
#
|
287
|
+
# For more information on streaming with Amazon Transcribe, see
|
288
|
+
# [Transcribing streaming audio][2].
|
289
|
+
#
|
290
|
+
#
|
291
|
+
#
|
292
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/call-analytics.html
|
293
|
+
# [2]: https://docs.aws.amazon.com/transcribe/latest/dg/streaming.html
|
294
|
+
#
|
295
|
+
# @option params [required, String] :language_code
|
296
|
+
# Specify the language code that represents the language spoken in your
|
297
|
+
# audio.
|
298
|
+
#
|
299
|
+
# If you're unsure of the language spoken in your audio, consider using
|
300
|
+
# `IdentifyLanguage` to enable automatic language identification.
|
301
|
+
#
|
302
|
+
# For a list of languages supported with streaming Call Analytics, refer
|
303
|
+
# to the [Supported languages][1] table.
|
304
|
+
#
|
305
|
+
#
|
306
|
+
#
|
307
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/supported-languages.html
|
308
|
+
#
|
309
|
+
# @option params [required, Integer] :media_sample_rate_hertz
|
310
|
+
# The sample rate of the input audio (in hertz). Low-quality audio, such
|
311
|
+
# as telephone audio, is typically around 8,000 Hz. High-quality audio
|
312
|
+
# typically ranges from 16,000 Hz to 48,000 Hz. Note that the sample
|
313
|
+
# rate you specify must match that of your audio.
|
314
|
+
#
|
315
|
+
# @option params [required, String] :media_encoding
|
316
|
+
# Specify the encoding of your input audio. Supported formats are:
|
317
|
+
#
|
318
|
+
# * FLAC
|
319
|
+
#
|
320
|
+
# * OPUS-encoded audio in an Ogg container
|
321
|
+
#
|
322
|
+
# * PCM (only signed 16-bit little-endian audio formats, which does not
|
323
|
+
# include WAV)
|
324
|
+
#
|
325
|
+
# For more information, see [Media formats][1].
|
326
|
+
#
|
327
|
+
#
|
328
|
+
#
|
329
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/how-input.html#how-input-audio
|
330
|
+
#
|
331
|
+
# @option params [String] :vocabulary_name
|
332
|
+
# Specify the name of the custom vocabulary that you want to use when
|
333
|
+
# processing your transcription. Note that vocabulary names are case
|
334
|
+
# sensitive.
|
335
|
+
#
|
336
|
+
# If the language of the specified custom vocabulary doesn't match the
|
337
|
+
# language identified in your media, the custom vocabulary is not
|
338
|
+
# applied to your transcription.
|
339
|
+
#
|
340
|
+
# For more information, see [Custom vocabularies][1].
|
341
|
+
#
|
342
|
+
#
|
343
|
+
#
|
344
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/custom-vocabulary.html
|
345
|
+
#
|
346
|
+
# @option params [String] :session_id
|
347
|
+
# Specify a name for your Call Analytics transcription session. If you
|
348
|
+
# don't include this parameter in your request, Amazon Transcribe
|
349
|
+
# generates an ID and returns it in the response.
|
350
|
+
#
|
351
|
+
# You can use a session ID to retry a streaming session.
|
352
|
+
#
|
353
|
+
# @option params [String] :vocabulary_filter_name
|
354
|
+
# Specify the name of the custom vocabulary filter that you want to use
|
355
|
+
# when processing your transcription. Note that vocabulary filter names
|
356
|
+
# are case sensitive.
|
357
|
+
#
|
358
|
+
# If the language of the specified custom vocabulary filter doesn't
|
359
|
+
# match the language identified in your media, the vocabulary filter is
|
360
|
+
# not applied to your transcription.
|
361
|
+
#
|
362
|
+
# For more information, see [Using vocabulary filtering with unwanted
|
363
|
+
# words][1].
|
364
|
+
#
|
365
|
+
#
|
366
|
+
#
|
367
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/vocabulary-filtering.html
|
368
|
+
#
|
369
|
+
# @option params [String] :vocabulary_filter_method
|
370
|
+
# Specify how you want your vocabulary filter applied to your
|
371
|
+
# transcript.
|
372
|
+
#
|
373
|
+
# To replace words with `***`, choose `mask`.
|
374
|
+
#
|
375
|
+
# To delete words, choose `remove`.
|
376
|
+
#
|
377
|
+
# To flag words without changing them, choose `tag`.
|
378
|
+
#
|
379
|
+
# @option params [String] :language_model_name
|
380
|
+
# Specify the name of the custom language model that you want to use
|
381
|
+
# when processing your transcription. Note that language model names are
|
382
|
+
# case sensitive.
|
383
|
+
#
|
384
|
+
# The language of the specified language model must match the language
|
385
|
+
# code you specify in your transcription request. If the languages
|
386
|
+
# don't match, the custom language model isn't applied. There are no
|
387
|
+
# errors or warnings associated with a language mismatch.
|
388
|
+
#
|
389
|
+
# For more information, see [Custom language models][1].
|
390
|
+
#
|
391
|
+
#
|
392
|
+
#
|
393
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/custom-language-models.html
|
394
|
+
#
|
395
|
+
# @option params [Boolean] :enable_partial_results_stabilization
|
396
|
+
# Enables partial result stabilization for your transcription. Partial
|
397
|
+
# result stabilization can reduce latency in your output, but may impact
|
398
|
+
# accuracy. For more information, see [Partial-result stabilization][1].
|
399
|
+
#
|
400
|
+
#
|
401
|
+
#
|
402
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/streaming.html#streaming-partial-result-stabilization
|
403
|
+
#
|
404
|
+
# @option params [String] :partial_results_stability
|
405
|
+
# Specify the level of stability to use when you enable partial results
|
406
|
+
# stabilization (`EnablePartialResultsStabilization`).
|
407
|
+
#
|
408
|
+
# Low stability provides the highest accuracy. High stability
|
409
|
+
# transcribes faster, but with slightly lower accuracy.
|
410
|
+
#
|
411
|
+
# For more information, see [Partial-result stabilization][1].
|
412
|
+
#
|
413
|
+
#
|
414
|
+
#
|
415
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/streaming.html#streaming-partial-result-stabilization
|
416
|
+
#
|
417
|
+
# @option params [String] :content_identification_type
|
418
|
+
# Labels all personally identifiable information (PII) identified in
|
419
|
+
# your transcript.
|
420
|
+
#
|
421
|
+
# Content identification is performed at the segment level; PII
|
422
|
+
# specified in `PiiEntityTypes` is flagged upon complete transcription
|
423
|
+
# of an audio segment.
|
424
|
+
#
|
425
|
+
# You can’t set `ContentIdentificationType` and `ContentRedactionType`
|
426
|
+
# in the same request. If you set both, your request returns a
|
427
|
+
# `BadRequestException`.
|
428
|
+
#
|
429
|
+
# For more information, see [Redacting or identifying personally
|
430
|
+
# identifiable information][1].
|
431
|
+
#
|
432
|
+
#
|
433
|
+
#
|
434
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/pii-redaction.html
|
435
|
+
#
|
436
|
+
# @option params [String] :content_redaction_type
|
437
|
+
# Redacts all personally identifiable information (PII) identified in
|
438
|
+
# your transcript.
|
439
|
+
#
|
440
|
+
# Content redaction is performed at the segment level; PII specified in
|
441
|
+
# `PiiEntityTypes` is redacted upon complete transcription of an audio
|
442
|
+
# segment.
|
443
|
+
#
|
444
|
+
# You can’t set `ContentRedactionType` and `ContentIdentificationType`
|
445
|
+
# in the same request. If you set both, your request returns a
|
446
|
+
# `BadRequestException`.
|
447
|
+
#
|
448
|
+
# For more information, see [Redacting or identifying personally
|
449
|
+
# identifiable information][1].
|
450
|
+
#
|
451
|
+
#
|
452
|
+
#
|
453
|
+
# [1]: https://docs.aws.amazon.com/transcribe/latest/dg/pii-redaction.html
|
454
|
+
#
|
455
|
+
# @option params [String] :pii_entity_types
|
456
|
+
# Specify which types of personally identifiable information (PII) you
|
457
|
+
# want to redact in your transcript. You can include as many types as
|
458
|
+
# you'd like, or you can select `ALL`.
|
459
|
+
#
|
460
|
+
# To include `PiiEntityTypes` in your Call Analytics request, you must
|
461
|
+
# also include either `ContentIdentificationType` or
|
462
|
+
# `ContentRedactionType`.
|
463
|
+
#
|
464
|
+
# Values must be comma-separated and can include: `BANK_ACCOUNT_NUMBER`,
|
465
|
+
# `BANK_ROUTING`, `CREDIT_DEBIT_NUMBER`, `CREDIT_DEBIT_CVV`,
|
466
|
+
# `CREDIT_DEBIT_EXPIRY`, `PIN`, `EMAIL`, `ADDRESS`, `NAME`, `PHONE`,
|
467
|
+
# `SSN`, or `ALL`.
|
468
|
+
#
|
469
|
+
# @return [Types::StartCallAnalyticsStreamTranscriptionResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
470
|
+
#
|
471
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#request_id #request_id} => String
|
472
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#language_code #language_code} => String
|
473
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#media_sample_rate_hertz #media_sample_rate_hertz} => Integer
|
474
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#media_encoding #media_encoding} => String
|
475
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#vocabulary_name #vocabulary_name} => String
|
476
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#session_id #session_id} => String
|
477
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#call_analytics_transcript_result_stream #call_analytics_transcript_result_stream} => Types::CallAnalyticsTranscriptResultStream
|
478
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#vocabulary_filter_name #vocabulary_filter_name} => String
|
479
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#vocabulary_filter_method #vocabulary_filter_method} => String
|
480
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#language_model_name #language_model_name} => String
|
481
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#enable_partial_results_stabilization #enable_partial_results_stabilization} => Boolean
|
482
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#partial_results_stability #partial_results_stability} => String
|
483
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#content_identification_type #content_identification_type} => String
|
484
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#content_redaction_type #content_redaction_type} => String
|
485
|
+
# * {Types::StartCallAnalyticsStreamTranscriptionResponse#pii_entity_types #pii_entity_types} => String
|
486
|
+
#
|
487
|
+
# @example Bi-directional EventStream Operation Example
|
488
|
+
#
|
489
|
+
# You can signal input events after initial request is
|
490
|
+
# established, events will be sent to stream
|
491
|
+
# immediately (once stream connection is established successfully).
|
492
|
+
#
|
493
|
+
# To signal events, you can call #signal methods from an Aws::TranscribeStreamingService::EventStreams::AudioStream object.
|
494
|
+
# Make sure signal events before calling #wait or #join! at async response.
|
495
|
+
#
|
496
|
+
# input_stream = Aws::TranscribeStreamingService::EventStreams::AudioStream.new
|
497
|
+
#
|
498
|
+
# async_resp = client.start_call_analytics_stream_transcription( # params input,
|
499
|
+
# input_event_stream_handler: input_stream) do |out_stream|
|
500
|
+
#
|
501
|
+
# # register callbacks for events arrival
|
502
|
+
# out_stream.on_utterance_event_event do |event|
|
503
|
+
# event # => Aws::TranscribeStreamingService::Types::UtteranceEvent
|
504
|
+
# end
|
505
|
+
# out_stream.on_category_event_event do |event|
|
506
|
+
# event # => Aws::TranscribeStreamingService::Types::CategoryEvent
|
507
|
+
# end
|
508
|
+
# out_stream.on_bad_request_exception_event do |event|
|
509
|
+
# event # => Aws::TranscribeStreamingService::Types::BadRequestException
|
510
|
+
# end
|
511
|
+
# out_stream.on_limit_exceeded_exception_event do |event|
|
512
|
+
# event # => Aws::TranscribeStreamingService::Types::LimitExceededException
|
513
|
+
# end
|
514
|
+
# out_stream.on_internal_failure_exception_event do |event|
|
515
|
+
# event # => Aws::TranscribeStreamingService::Types::InternalFailureException
|
516
|
+
# end
|
517
|
+
# out_stream.on_conflict_exception_event do |event|
|
518
|
+
# event # => Aws::TranscribeStreamingService::Types::ConflictException
|
519
|
+
# end
|
520
|
+
# out_stream.on_service_unavailable_exception_event do |event|
|
521
|
+
# event # => Aws::TranscribeStreamingService::Types::ServiceUnavailableException
|
522
|
+
# end
|
523
|
+
#
|
524
|
+
# end
|
525
|
+
# # => returns Aws::Seahorse::Client::AsyncResponse
|
526
|
+
#
|
527
|
+
# # signal events
|
528
|
+
# input_stream.signal_audio_event_event( ... )
|
529
|
+
# input_stream.signal_configuration_event_event( ... )
|
530
|
+
#
|
531
|
+
# # make sure signaling :end_stream in the end
|
532
|
+
# input_stream.signal_end_stream
|
533
|
+
#
|
534
|
+
# # wait until stream is closed before finalizing sync response
|
535
|
+
# resp = async_resp.wait
|
536
|
+
# # Or close stream and finalizing sync response immediately
|
537
|
+
# # resp = async_resp.join!
|
538
|
+
#
|
539
|
+
# Inorder to streamingly processing events received, you can also provide an Aws::TranscribeStreamingService::EventStreams::CallAnalyticsTranscriptResultStream
|
540
|
+
# object to register callbacks before initializing request instead of processing from request block
|
541
|
+
#
|
542
|
+
# output_stream = Aws::TranscribeStreamingService::EventStreams::CallAnalyticsTranscriptResultStream.new
|
543
|
+
# # register callbacks for events arrival
|
544
|
+
# output_stream.on_utterance_event_event do |event|
|
545
|
+
# event # => Aws::TranscribeStreamingService::Types::UtteranceEvent
|
546
|
+
# end
|
547
|
+
# output_stream.on_category_event_event do |event|
|
548
|
+
# event # => Aws::TranscribeStreamingService::Types::CategoryEvent
|
549
|
+
# end
|
550
|
+
# output_stream.on_bad_request_exception_event do |event|
|
551
|
+
# event # => Aws::TranscribeStreamingService::Types::BadRequestException
|
552
|
+
# end
|
553
|
+
# output_stream.on_limit_exceeded_exception_event do |event|
|
554
|
+
# event # => Aws::TranscribeStreamingService::Types::LimitExceededException
|
555
|
+
# end
|
556
|
+
# output_stream.on_internal_failure_exception_event do |event|
|
557
|
+
# event # => Aws::TranscribeStreamingService::Types::InternalFailureException
|
558
|
+
# end
|
559
|
+
# output_stream.on_conflict_exception_event do |event|
|
560
|
+
# event # => Aws::TranscribeStreamingService::Types::ConflictException
|
561
|
+
# end
|
562
|
+
# output_stream.on_service_unavailable_exception_event do |event|
|
563
|
+
# event # => Aws::TranscribeStreamingService::Types::ServiceUnavailableException
|
564
|
+
# end
|
565
|
+
# output_stream.on_error_event do |event|
|
566
|
+
# # catch unmodeled error event in the stream
|
567
|
+
# raise event
|
568
|
+
# # => Aws::Errors::EventError
|
569
|
+
# # event.event_type => :error
|
570
|
+
# # event.error_code => String
|
571
|
+
# # event.error_message => String
|
572
|
+
# end
|
573
|
+
#
|
574
|
+
# async_resp = client.start_call_analytics_stream_transcription ( #params input,
|
575
|
+
# input_event_stream_handler: input_stream
|
576
|
+
# output_event_stream_handler: output_stream
|
577
|
+
# )
|
578
|
+
#
|
579
|
+
# resp = async_resp.wait!
|
580
|
+
#
|
581
|
+
# Besides above usage patterns for process events when they arrive immediately, you can also
|
582
|
+
# iterate through events after response complete.
|
583
|
+
#
|
584
|
+
# Events are available at resp.call_analytics_transcript_result_stream # => Enumerator
|
585
|
+
#
|
586
|
+
# @example Request syntax with placeholder values
|
587
|
+
#
|
588
|
+
# async_resp = async_client.start_call_analytics_stream_transcription({
|
589
|
+
# language_code: "en-US", # required, accepts en-US, en-GB, es-US, fr-CA, fr-FR, en-AU, it-IT, de-DE, pt-BR
|
590
|
+
# media_sample_rate_hertz: 1, # required
|
591
|
+
# media_encoding: "pcm", # required, accepts pcm, ogg-opus, flac
|
592
|
+
# vocabulary_name: "VocabularyName",
|
593
|
+
# session_id: "SessionId",
|
594
|
+
# input_event_stream_hander: EventStreams::AudioStream.new,
|
595
|
+
# vocabulary_filter_name: "VocabularyFilterName",
|
596
|
+
# vocabulary_filter_method: "remove", # accepts remove, mask, tag
|
597
|
+
# language_model_name: "ModelName",
|
598
|
+
# enable_partial_results_stabilization: false,
|
599
|
+
# partial_results_stability: "high", # accepts high, medium, low
|
600
|
+
# content_identification_type: "PII", # accepts PII
|
601
|
+
# content_redaction_type: "PII", # accepts PII
|
602
|
+
# pii_entity_types: "PiiEntityTypes",
|
603
|
+
# })
|
604
|
+
# # => Seahorse::Client::AsyncResponse
|
605
|
+
# async_resp.wait
|
606
|
+
# # => Seahorse::Client::Response
|
607
|
+
# # Or use async_resp.join!
|
608
|
+
#
|
609
|
+
# @example Response structure
|
610
|
+
#
|
611
|
+
# resp.request_id #=> String
|
612
|
+
# resp.language_code #=> String, one of "en-US", "en-GB", "es-US", "fr-CA", "fr-FR", "en-AU", "it-IT", "de-DE", "pt-BR"
|
613
|
+
# resp.media_sample_rate_hertz #=> Integer
|
614
|
+
# resp.media_encoding #=> String, one of "pcm", "ogg-opus", "flac"
|
615
|
+
# resp.vocabulary_name #=> String
|
616
|
+
# resp.session_id #=> String
|
617
|
+
# All events are available at resp.call_analytics_transcript_result_stream:
|
618
|
+
# resp.call_analytics_transcript_result_stream #=> Enumerator
|
619
|
+
# resp.call_analytics_transcript_result_stream.event_types #=> [:utterance_event, :category_event, :bad_request_exception, :limit_exceeded_exception, :internal_failure_exception, :conflict_exception, :service_unavailable_exception]
|
620
|
+
#
|
621
|
+
# For :utterance_event event available at #on_utterance_event_event callback and response eventstream enumerator:
|
622
|
+
# event.utterance_id #=> String
|
623
|
+
# event.is_partial #=> Boolean
|
624
|
+
# event.participant_role #=> String, one of "AGENT", "CUSTOMER"
|
625
|
+
# event.begin_offset_millis #=> Integer
|
626
|
+
# event.end_offset_millis #=> Integer
|
627
|
+
# event.transcript #=> String
|
628
|
+
# event.items #=> Array
|
629
|
+
# event.items[0].begin_offset_millis #=> Integer
|
630
|
+
# event.items[0].end_offset_millis #=> Integer
|
631
|
+
# event.items[0].type #=> String, one of "pronunciation", "punctuation"
|
632
|
+
# event.items[0].content #=> String
|
633
|
+
# event.items[0].confidence #=> Float
|
634
|
+
# event.items[0].vocabulary_filter_match #=> Boolean
|
635
|
+
# event.items[0].stable #=> Boolean
|
636
|
+
# event.entities #=> Array
|
637
|
+
# event.entities[0].begin_offset_millis #=> Integer
|
638
|
+
# event.entities[0].end_offset_millis #=> Integer
|
639
|
+
# event.entities[0].category #=> String
|
640
|
+
# event.entities[0].type #=> String
|
641
|
+
# event.entities[0].content #=> String
|
642
|
+
# event.entities[0].confidence #=> Float
|
643
|
+
# event.sentiment #=> String, one of "POSITIVE", "NEGATIVE", "MIXED", "NEUTRAL"
|
644
|
+
# event.issues_detected #=> Array
|
645
|
+
# event.issues_detected[0].character_offsets.begin #=> Integer
|
646
|
+
# event.issues_detected[0].character_offsets.end #=> Integer
|
647
|
+
#
|
648
|
+
# For :category_event event available at #on_category_event_event callback and response eventstream enumerator:
|
649
|
+
# event.matched_categories #=> Array
|
650
|
+
# event.matched_categories[0] #=> String
|
651
|
+
# event.matched_details #=> Hash
|
652
|
+
# event.matched_details["String"].timestamp_ranges #=> Array
|
653
|
+
# event.matched_details["String"].timestamp_ranges[0].begin_offset_millis #=> Integer
|
654
|
+
# event.matched_details["String"].timestamp_ranges[0].end_offset_millis #=> Integer
|
655
|
+
#
|
656
|
+
# For :bad_request_exception event available at #on_bad_request_exception_event callback and response eventstream enumerator:
|
657
|
+
# event.message #=> String
|
658
|
+
#
|
659
|
+
# For :limit_exceeded_exception event available at #on_limit_exceeded_exception_event callback and response eventstream enumerator:
|
660
|
+
# event.message #=> String
|
661
|
+
#
|
662
|
+
# For :internal_failure_exception event available at #on_internal_failure_exception_event callback and response eventstream enumerator:
|
663
|
+
# event.message #=> String
|
664
|
+
#
|
665
|
+
# For :conflict_exception event available at #on_conflict_exception_event callback and response eventstream enumerator:
|
666
|
+
# event.message #=> String
|
667
|
+
#
|
668
|
+
# For :service_unavailable_exception event available at #on_service_unavailable_exception_event callback and response eventstream enumerator:
|
669
|
+
# event.message #=> String
|
670
|
+
#
|
671
|
+
# resp.vocabulary_filter_name #=> String
|
672
|
+
# resp.vocabulary_filter_method #=> String, one of "remove", "mask", "tag"
|
673
|
+
# resp.language_model_name #=> String
|
674
|
+
# resp.enable_partial_results_stabilization #=> Boolean
|
675
|
+
# resp.partial_results_stability #=> String, one of "high", "medium", "low"
|
676
|
+
# resp.content_identification_type #=> String, one of "PII"
|
677
|
+
# resp.content_redaction_type #=> String, one of "PII"
|
678
|
+
# resp.pii_entity_types #=> String
|
679
|
+
#
|
680
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/transcribe-streaming-2017-10-26/StartCallAnalyticsStreamTranscription AWS API Documentation
|
681
|
+
#
|
682
|
+
# @overload start_call_analytics_stream_transcription(params = {})
|
683
|
+
# @param [Hash] params ({})
|
684
|
+
def start_call_analytics_stream_transcription(params = {}, options = {}, &block)
|
685
|
+
params = params.dup
|
686
|
+
input_event_stream_handler = _event_stream_handler(
|
687
|
+
:input,
|
688
|
+
params.delete(:input_event_stream_handler),
|
689
|
+
EventStreams::AudioStream
|
690
|
+
)
|
691
|
+
output_event_stream_handler = _event_stream_handler(
|
692
|
+
:output,
|
693
|
+
params.delete(:output_event_stream_handler) || params.delete(:event_stream_handler),
|
694
|
+
EventStreams::CallAnalyticsTranscriptResultStream
|
695
|
+
)
|
696
|
+
|
697
|
+
yield(output_event_stream_handler) if block_given?
|
698
|
+
|
699
|
+
req = build_request(:start_call_analytics_stream_transcription, params)
|
700
|
+
|
701
|
+
req.context[:input_event_stream_handler] = input_event_stream_handler
|
702
|
+
req.handlers.add(Aws::Binary::EncodeHandler, priority: 55)
|
703
|
+
req.context[:output_event_stream_handler] = output_event_stream_handler
|
704
|
+
req.handlers.add(Aws::Binary::DecodeHandler, priority: 55)
|
705
|
+
|
706
|
+
req.send_request(options, &block)
|
707
|
+
end
|
708
|
+
|
274
709
|
# Starts a bidirectional HTTP/2 or WebSocket stream where audio is
|
275
710
|
# streamed to Amazon Transcribe Medical and the transcription results
|
276
711
|
# are streamed to your application.
|
277
712
|
#
|
713
|
+
# The following parameters are required:
|
714
|
+
#
|
715
|
+
# * `language-code`
|
716
|
+
#
|
717
|
+
# * `media-encoding`
|
718
|
+
#
|
719
|
+
# * `sample-rate`
|
720
|
+
#
|
278
721
|
# For more information on streaming with Amazon Transcribe Medical, see
|
279
722
|
# [Transcribing streaming audio][1].
|
280
723
|
#
|
@@ -430,6 +873,7 @@ module Aws::TranscribeStreamingService
|
|
430
873
|
#
|
431
874
|
# # signal events
|
432
875
|
# input_stream.signal_audio_event_event( ... )
|
876
|
+
# input_stream.signal_configuration_event_event( ... )
|
433
877
|
#
|
434
878
|
# # make sure signaling :end_stream in the end
|
435
879
|
# input_stream.signal_end_stream
|
@@ -594,15 +1038,13 @@ module Aws::TranscribeStreamingService
|
|
594
1038
|
# streamed to Amazon Transcribe and the transcription results are
|
595
1039
|
# streamed to your application.
|
596
1040
|
#
|
597
|
-
# The following are
|
598
|
-
#
|
599
|
-
# * language-code
|
1041
|
+
# The following parameters are required:
|
600
1042
|
#
|
601
|
-
# *
|
1043
|
+
# * `language-code` or `identify-language`
|
602
1044
|
#
|
603
|
-
# *
|
1045
|
+
# * `media-encoding`
|
604
1046
|
#
|
605
|
-
# *
|
1047
|
+
# * `sample-rate`
|
606
1048
|
#
|
607
1049
|
# For more information on streaming with Amazon Transcribe, see
|
608
1050
|
# [Transcribing streaming audio][1].
|
@@ -632,7 +1074,7 @@ module Aws::TranscribeStreamingService
|
|
632
1074
|
# rate you specify must match that of your audio.
|
633
1075
|
#
|
634
1076
|
# @option params [required, String] :media_encoding
|
635
|
-
# Specify the encoding
|
1077
|
+
# Specify the encoding of your input audio. Supported formats are:
|
636
1078
|
#
|
637
1079
|
# * FLAC
|
638
1080
|
#
|
@@ -653,7 +1095,8 @@ module Aws::TranscribeStreamingService
|
|
653
1095
|
# sensitive.
|
654
1096
|
#
|
655
1097
|
# If the language of the specified custom vocabulary doesn't match the
|
656
|
-
# language identified in your media,
|
1098
|
+
# language identified in your media, the custom vocabulary is not
|
1099
|
+
# applied to your transcription.
|
657
1100
|
#
|
658
1101
|
# This parameter is **not** intended for use with the `IdentifyLanguage`
|
659
1102
|
# parameter. If you're including `IdentifyLanguage` in your request and
|
@@ -679,7 +1122,8 @@ module Aws::TranscribeStreamingService
|
|
679
1122
|
# are case sensitive.
|
680
1123
|
#
|
681
1124
|
# If the language of the specified custom vocabulary filter doesn't
|
682
|
-
# match the language identified in your media,
|
1125
|
+
# match the language identified in your media, the vocabulary filter is
|
1126
|
+
# not applied to your transcription.
|
683
1127
|
#
|
684
1128
|
# This parameter is **not** intended for use with the `IdentifyLanguage`
|
685
1129
|
# parameter. If you're including `IdentifyLanguage` in your request and
|
@@ -815,8 +1259,8 @@ module Aws::TranscribeStreamingService
|
|
815
1259
|
#
|
816
1260
|
# The language of the specified language model must match the language
|
817
1261
|
# code you specify in your transcription request. If the languages
|
818
|
-
# don't match, the language model isn't applied. There are no
|
819
|
-
# or warnings associated with a language mismatch.
|
1262
|
+
# don't match, the custom language model isn't applied. There are no
|
1263
|
+
# errors or warnings associated with a language mismatch.
|
820
1264
|
#
|
821
1265
|
# For more information, see [Custom language models][1].
|
822
1266
|
#
|
@@ -981,6 +1425,7 @@ module Aws::TranscribeStreamingService
|
|
981
1425
|
#
|
982
1426
|
# # signal events
|
983
1427
|
# input_stream.signal_audio_event_event( ... )
|
1428
|
+
# input_stream.signal_configuration_event_event( ... )
|
984
1429
|
#
|
985
1430
|
# # make sure signaling :end_stream in the end
|
986
1431
|
# input_stream.signal_end_stream
|
@@ -1182,7 +1627,7 @@ module Aws::TranscribeStreamingService
|
|
1182
1627
|
http_response: Seahorse::Client::Http::AsyncResponse.new,
|
1183
1628
|
config: config)
|
1184
1629
|
context[:gem_name] = 'aws-sdk-transcribestreamingservice'
|
1185
|
-
context[:gem_version] = '1.
|
1630
|
+
context[:gem_version] = '1.46.0'
|
1186
1631
|
Seahorse::Client::Request.new(handlers, context)
|
1187
1632
|
end
|
1188
1633
|
|
@@ -392,7 +392,7 @@ module Aws::TranscribeStreamingService
|
|
392
392
|
params: params,
|
393
393
|
config: config)
|
394
394
|
context[:gem_name] = 'aws-sdk-transcribestreamingservice'
|
395
|
-
context[:gem_version] = '1.
|
395
|
+
context[:gem_version] = '1.46.0'
|
396
396
|
Seahorse::Client::Request.new(handlers, context)
|
397
397
|
end
|
398
398
|
|