aws-sdk-qconnect 1.46.0 → 1.48.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +10 -0
- data/VERSION +1 -1
- data/lib/aws-sdk-qconnect/client.rb +25 -29
- data/lib/aws-sdk-qconnect/client_api.rb +5 -12
- data/lib/aws-sdk-qconnect/types.rb +19 -42
- data/lib/aws-sdk-qconnect.rb +1 -1
- data/sig/client.rbs +8 -12
- data/sig/types.rbs +5 -16
- metadata +3 -3
checksums.yaml
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
---
|
|
2
2
|
SHA256:
|
|
3
|
-
metadata.gz:
|
|
4
|
-
data.tar.gz:
|
|
3
|
+
metadata.gz: 737e4ee1ca02ba610fb9d4d7b29439b4092b957e2d7a080d8407bf359c690560
|
|
4
|
+
data.tar.gz: eacec788d0fb123ffe936f7cab221b91357a808013cd245ce816b8b94023c9cd
|
|
5
5
|
SHA512:
|
|
6
|
-
metadata.gz:
|
|
7
|
-
data.tar.gz:
|
|
6
|
+
metadata.gz: 165120e837d8ade4a3fbef6e60bb36aef12066703dba30b99adb31ed4b0a232b369d5239368ce6c9c84844e395c89e266e04589f653d0b83b4a233228934a0c5
|
|
7
|
+
data.tar.gz: 9202e10c27266a6d7fb6944dc2ed621b843c3d70590199db1f506ff6534bd2ede4508d5e85c4619cf07aa8ad3c4878e135f97b3ed4a5f1f6d1094bbb63cd2456
|
data/CHANGELOG.md
CHANGED
|
@@ -1,6 +1,16 @@
|
|
|
1
1
|
Unreleased Changes
|
|
2
2
|
------------------
|
|
3
3
|
|
|
4
|
+
1.48.0 (2026-01-16)
|
|
5
|
+
------------------
|
|
6
|
+
|
|
7
|
+
* Feature - Code Generated Changes, see `./build_tools` or `aws-sdk-core`'s CHANGELOG.md for details.
|
|
8
|
+
|
|
9
|
+
1.47.0 (2026-01-15)
|
|
10
|
+
------------------
|
|
11
|
+
|
|
12
|
+
* Feature - Fix inference configuration shapes for the CreateAIPrompt and UpdateAIPrompt APIs, Modify Text Length Limit for SendMessage API
|
|
13
|
+
|
|
4
14
|
1.46.0 (2026-01-08)
|
|
5
15
|
------------------
|
|
6
16
|
|
data/VERSION
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
1.
|
|
1
|
+
1.48.0
|
|
@@ -1574,12 +1574,10 @@ module Aws::QConnect
|
|
|
1574
1574
|
# },
|
|
1575
1575
|
# description: "Description",
|
|
1576
1576
|
# inference_configuration: {
|
|
1577
|
-
#
|
|
1578
|
-
#
|
|
1579
|
-
#
|
|
1580
|
-
#
|
|
1581
|
-
# max_tokens_to_sample: 1,
|
|
1582
|
-
# },
|
|
1577
|
+
# temperature: 1.0,
|
|
1578
|
+
# top_p: 1.0,
|
|
1579
|
+
# top_k: 1,
|
|
1580
|
+
# max_tokens_to_sample: 1,
|
|
1583
1581
|
# },
|
|
1584
1582
|
# })
|
|
1585
1583
|
#
|
|
@@ -1595,10 +1593,10 @@ module Aws::QConnect
|
|
|
1595
1593
|
# resp.ai_prompt.model_id #=> String
|
|
1596
1594
|
# resp.ai_prompt.api_format #=> String, one of "ANTHROPIC_CLAUDE_MESSAGES", "ANTHROPIC_CLAUDE_TEXT_COMPLETIONS", "MESSAGES", "TEXT_COMPLETIONS"
|
|
1597
1595
|
# resp.ai_prompt.template_configuration.text_full_ai_prompt_edit_template_configuration.text #=> String
|
|
1598
|
-
# resp.ai_prompt.inference_configuration.
|
|
1599
|
-
# resp.ai_prompt.inference_configuration.
|
|
1600
|
-
# resp.ai_prompt.inference_configuration.
|
|
1601
|
-
# resp.ai_prompt.inference_configuration.
|
|
1596
|
+
# resp.ai_prompt.inference_configuration.temperature #=> Float
|
|
1597
|
+
# resp.ai_prompt.inference_configuration.top_p #=> Float
|
|
1598
|
+
# resp.ai_prompt.inference_configuration.top_k #=> Integer
|
|
1599
|
+
# resp.ai_prompt.inference_configuration.max_tokens_to_sample #=> Integer
|
|
1602
1600
|
# resp.ai_prompt.modified_time #=> Time
|
|
1603
1601
|
# resp.ai_prompt.description #=> String
|
|
1604
1602
|
# resp.ai_prompt.visibility_status #=> String, one of "SAVED", "PUBLISHED"
|
|
@@ -1667,10 +1665,10 @@ module Aws::QConnect
|
|
|
1667
1665
|
# resp.ai_prompt.model_id #=> String
|
|
1668
1666
|
# resp.ai_prompt.api_format #=> String, one of "ANTHROPIC_CLAUDE_MESSAGES", "ANTHROPIC_CLAUDE_TEXT_COMPLETIONS", "MESSAGES", "TEXT_COMPLETIONS"
|
|
1669
1667
|
# resp.ai_prompt.template_configuration.text_full_ai_prompt_edit_template_configuration.text #=> String
|
|
1670
|
-
# resp.ai_prompt.inference_configuration.
|
|
1671
|
-
# resp.ai_prompt.inference_configuration.
|
|
1672
|
-
# resp.ai_prompt.inference_configuration.
|
|
1673
|
-
# resp.ai_prompt.inference_configuration.
|
|
1668
|
+
# resp.ai_prompt.inference_configuration.temperature #=> Float
|
|
1669
|
+
# resp.ai_prompt.inference_configuration.top_p #=> Float
|
|
1670
|
+
# resp.ai_prompt.inference_configuration.top_k #=> Integer
|
|
1671
|
+
# resp.ai_prompt.inference_configuration.max_tokens_to_sample #=> Integer
|
|
1674
1672
|
# resp.ai_prompt.modified_time #=> Time
|
|
1675
1673
|
# resp.ai_prompt.description #=> String
|
|
1676
1674
|
# resp.ai_prompt.visibility_status #=> String, one of "SAVED", "PUBLISHED"
|
|
@@ -3992,10 +3990,10 @@ module Aws::QConnect
|
|
|
3992
3990
|
# resp.ai_prompt.model_id #=> String
|
|
3993
3991
|
# resp.ai_prompt.api_format #=> String, one of "ANTHROPIC_CLAUDE_MESSAGES", "ANTHROPIC_CLAUDE_TEXT_COMPLETIONS", "MESSAGES", "TEXT_COMPLETIONS"
|
|
3994
3992
|
# resp.ai_prompt.template_configuration.text_full_ai_prompt_edit_template_configuration.text #=> String
|
|
3995
|
-
# resp.ai_prompt.inference_configuration.
|
|
3996
|
-
# resp.ai_prompt.inference_configuration.
|
|
3997
|
-
# resp.ai_prompt.inference_configuration.
|
|
3998
|
-
# resp.ai_prompt.inference_configuration.
|
|
3993
|
+
# resp.ai_prompt.inference_configuration.temperature #=> Float
|
|
3994
|
+
# resp.ai_prompt.inference_configuration.top_p #=> Float
|
|
3995
|
+
# resp.ai_prompt.inference_configuration.top_k #=> Integer
|
|
3996
|
+
# resp.ai_prompt.inference_configuration.max_tokens_to_sample #=> Integer
|
|
3999
3997
|
# resp.ai_prompt.modified_time #=> Time
|
|
4000
3998
|
# resp.ai_prompt.description #=> String
|
|
4001
3999
|
# resp.ai_prompt.visibility_status #=> String, one of "SAVED", "PUBLISHED"
|
|
@@ -8317,12 +8315,10 @@ module Aws::QConnect
|
|
|
8317
8315
|
# description: "Description",
|
|
8318
8316
|
# model_id: "AIPromptModelIdentifier",
|
|
8319
8317
|
# inference_configuration: {
|
|
8320
|
-
#
|
|
8321
|
-
#
|
|
8322
|
-
#
|
|
8323
|
-
#
|
|
8324
|
-
# max_tokens_to_sample: 1,
|
|
8325
|
-
# },
|
|
8318
|
+
# temperature: 1.0,
|
|
8319
|
+
# top_p: 1.0,
|
|
8320
|
+
# top_k: 1,
|
|
8321
|
+
# max_tokens_to_sample: 1,
|
|
8326
8322
|
# },
|
|
8327
8323
|
# })
|
|
8328
8324
|
#
|
|
@@ -8338,10 +8334,10 @@ module Aws::QConnect
|
|
|
8338
8334
|
# resp.ai_prompt.model_id #=> String
|
|
8339
8335
|
# resp.ai_prompt.api_format #=> String, one of "ANTHROPIC_CLAUDE_MESSAGES", "ANTHROPIC_CLAUDE_TEXT_COMPLETIONS", "MESSAGES", "TEXT_COMPLETIONS"
|
|
8340
8336
|
# resp.ai_prompt.template_configuration.text_full_ai_prompt_edit_template_configuration.text #=> String
|
|
8341
|
-
# resp.ai_prompt.inference_configuration.
|
|
8342
|
-
# resp.ai_prompt.inference_configuration.
|
|
8343
|
-
# resp.ai_prompt.inference_configuration.
|
|
8344
|
-
# resp.ai_prompt.inference_configuration.
|
|
8337
|
+
# resp.ai_prompt.inference_configuration.temperature #=> Float
|
|
8338
|
+
# resp.ai_prompt.inference_configuration.top_p #=> Float
|
|
8339
|
+
# resp.ai_prompt.inference_configuration.top_k #=> Integer
|
|
8340
|
+
# resp.ai_prompt.inference_configuration.max_tokens_to_sample #=> Integer
|
|
8345
8341
|
# resp.ai_prompt.modified_time #=> Time
|
|
8346
8342
|
# resp.ai_prompt.description #=> String
|
|
8347
8343
|
# resp.ai_prompt.visibility_status #=> String, one of "SAVED", "PUBLISHED"
|
|
@@ -9463,7 +9459,7 @@ module Aws::QConnect
|
|
|
9463
9459
|
tracer: tracer
|
|
9464
9460
|
)
|
|
9465
9461
|
context[:gem_name] = 'aws-sdk-qconnect'
|
|
9466
|
-
context[:gem_version] = '1.
|
|
9462
|
+
context[:gem_version] = '1.48.0'
|
|
9467
9463
|
Seahorse::Client::Request.new(handlers, context)
|
|
9468
9464
|
end
|
|
9469
9465
|
|
|
@@ -39,7 +39,7 @@ module Aws::QConnect
|
|
|
39
39
|
AIGuardrailWordPolicyConfig = Shapes::StructureShape.new(name: 'AIGuardrailWordPolicyConfig')
|
|
40
40
|
AIPromptAPIFormat = Shapes::StringShape.new(name: 'AIPromptAPIFormat')
|
|
41
41
|
AIPromptData = Shapes::StructureShape.new(name: 'AIPromptData')
|
|
42
|
-
AIPromptInferenceConfiguration = Shapes::
|
|
42
|
+
AIPromptInferenceConfiguration = Shapes::StructureShape.new(name: 'AIPromptInferenceConfiguration')
|
|
43
43
|
AIPromptModelIdentifier = Shapes::StringShape.new(name: 'AIPromptModelIdentifier')
|
|
44
44
|
AIPromptSummary = Shapes::StructureShape.new(name: 'AIPromptSummary')
|
|
45
45
|
AIPromptSummaryList = Shapes::ListShape.new(name: 'AIPromptSummaryList')
|
|
@@ -587,7 +587,6 @@ module Aws::QConnect
|
|
|
587
587
|
Tags = Shapes::MapShape.new(name: 'Tags')
|
|
588
588
|
TargetType = Shapes::StringShape.new(name: 'TargetType')
|
|
589
589
|
TextAIPrompt = Shapes::StringShape.new(name: 'TextAIPrompt')
|
|
590
|
-
TextAIPromptInferenceConfiguration = Shapes::StructureShape.new(name: 'TextAIPromptInferenceConfiguration')
|
|
591
590
|
TextData = Shapes::StructureShape.new(name: 'TextData')
|
|
592
591
|
TextFullAIPromptEditTemplateConfiguration = Shapes::StructureShape.new(name: 'TextFullAIPromptEditTemplateConfiguration')
|
|
593
592
|
TextMessage = Shapes::StructureShape.new(name: 'TextMessage')
|
|
@@ -814,10 +813,10 @@ module Aws::QConnect
|
|
|
814
813
|
AIPromptData.add_member(:status, Shapes::ShapeRef.new(shape: Status, location_name: "status"))
|
|
815
814
|
AIPromptData.struct_class = Types::AIPromptData
|
|
816
815
|
|
|
817
|
-
AIPromptInferenceConfiguration.add_member(:
|
|
818
|
-
AIPromptInferenceConfiguration.add_member(:
|
|
819
|
-
AIPromptInferenceConfiguration.
|
|
820
|
-
AIPromptInferenceConfiguration.
|
|
816
|
+
AIPromptInferenceConfiguration.add_member(:temperature, Shapes::ShapeRef.new(shape: Probability, location_name: "temperature"))
|
|
817
|
+
AIPromptInferenceConfiguration.add_member(:top_p, Shapes::ShapeRef.new(shape: Probability, location_name: "topP"))
|
|
818
|
+
AIPromptInferenceConfiguration.add_member(:top_k, Shapes::ShapeRef.new(shape: TopK, location_name: "topK"))
|
|
819
|
+
AIPromptInferenceConfiguration.add_member(:max_tokens_to_sample, Shapes::ShapeRef.new(shape: MaxTokensToSample, location_name: "maxTokensToSample"))
|
|
821
820
|
AIPromptInferenceConfiguration.struct_class = Types::AIPromptInferenceConfiguration
|
|
822
821
|
|
|
823
822
|
AIPromptSummary.add_member(:name, Shapes::ShapeRef.new(shape: Name, required: true, location_name: "name"))
|
|
@@ -2998,12 +2997,6 @@ module Aws::QConnect
|
|
|
2998
2997
|
Tags.key = Shapes::ShapeRef.new(shape: TagKey)
|
|
2999
2998
|
Tags.value = Shapes::ShapeRef.new(shape: TagValue)
|
|
3000
2999
|
|
|
3001
|
-
TextAIPromptInferenceConfiguration.add_member(:temperature, Shapes::ShapeRef.new(shape: Probability, location_name: "temperature"))
|
|
3002
|
-
TextAIPromptInferenceConfiguration.add_member(:top_p, Shapes::ShapeRef.new(shape: Probability, location_name: "topP"))
|
|
3003
|
-
TextAIPromptInferenceConfiguration.add_member(:top_k, Shapes::ShapeRef.new(shape: TopK, location_name: "topK"))
|
|
3004
|
-
TextAIPromptInferenceConfiguration.add_member(:max_tokens_to_sample, Shapes::ShapeRef.new(shape: MaxTokensToSample, location_name: "maxTokensToSample"))
|
|
3005
|
-
TextAIPromptInferenceConfiguration.struct_class = Types::TextAIPromptInferenceConfiguration
|
|
3006
|
-
|
|
3007
3000
|
TextData.add_member(:title, Shapes::ShapeRef.new(shape: DocumentText, location_name: "title"))
|
|
3008
3001
|
TextData.add_member(:excerpt, Shapes::ShapeRef.new(shape: DocumentText, location_name: "excerpt"))
|
|
3009
3002
|
TextData.struct_class = Types::TextData
|
|
@@ -692,25 +692,32 @@ module Aws::QConnect
|
|
|
692
692
|
|
|
693
693
|
# The configuration for inference parameters when using AI Prompts.
|
|
694
694
|
#
|
|
695
|
-
#
|
|
695
|
+
# @!attribute [rw] temperature
|
|
696
|
+
# The temperature setting for controlling randomness in the generated
|
|
697
|
+
# response.
|
|
698
|
+
# @return [Float]
|
|
696
699
|
#
|
|
697
|
-
#
|
|
700
|
+
# @!attribute [rw] top_p
|
|
701
|
+
# The top-P sampling parameter for nucleus sampling.
|
|
702
|
+
# @return [Float]
|
|
698
703
|
#
|
|
699
|
-
# @!attribute [rw]
|
|
700
|
-
# The
|
|
701
|
-
# @return [
|
|
704
|
+
# @!attribute [rw] top_k
|
|
705
|
+
# The top-K sampling parameter for token selection.
|
|
706
|
+
# @return [Integer]
|
|
707
|
+
#
|
|
708
|
+
# @!attribute [rw] max_tokens_to_sample
|
|
709
|
+
# The maximum number of tokens to generate in the response.
|
|
710
|
+
# @return [Integer]
|
|
702
711
|
#
|
|
703
712
|
# @see http://docs.aws.amazon.com/goto/WebAPI/qconnect-2020-10-19/AIPromptInferenceConfiguration AWS API Documentation
|
|
704
713
|
#
|
|
705
714
|
class AIPromptInferenceConfiguration < Struct.new(
|
|
706
|
-
:
|
|
707
|
-
:
|
|
708
|
-
|
|
715
|
+
:temperature,
|
|
716
|
+
:top_p,
|
|
717
|
+
:top_k,
|
|
718
|
+
:max_tokens_to_sample)
|
|
719
|
+
SENSITIVE = [:temperature, :top_p, :top_k, :max_tokens_to_sample]
|
|
709
720
|
include Aws::Structure
|
|
710
|
-
include Aws::Structure::Union
|
|
711
|
-
|
|
712
|
-
class TextAiPromptInferenceConfiguration < AIPromptInferenceConfiguration; end
|
|
713
|
-
class Unknown < AIPromptInferenceConfiguration; end
|
|
714
721
|
end
|
|
715
722
|
|
|
716
723
|
# The summary of the AI Prompt.
|
|
@@ -11279,36 +11286,6 @@ module Aws::QConnect
|
|
|
11279
11286
|
#
|
|
11280
11287
|
class TagResourceResponse < Aws::EmptyStructure; end
|
|
11281
11288
|
|
|
11282
|
-
# Inference configuration for text-based AI Prompts.
|
|
11283
|
-
#
|
|
11284
|
-
# @!attribute [rw] temperature
|
|
11285
|
-
# The temperature setting for controlling randomness in the generated
|
|
11286
|
-
# response.
|
|
11287
|
-
# @return [Float]
|
|
11288
|
-
#
|
|
11289
|
-
# @!attribute [rw] top_p
|
|
11290
|
-
# The top-P sampling parameter for nucleus sampling.
|
|
11291
|
-
# @return [Float]
|
|
11292
|
-
#
|
|
11293
|
-
# @!attribute [rw] top_k
|
|
11294
|
-
# The top-K sampling parameter for token selection.
|
|
11295
|
-
# @return [Integer]
|
|
11296
|
-
#
|
|
11297
|
-
# @!attribute [rw] max_tokens_to_sample
|
|
11298
|
-
# The maximum number of tokens to generate in the response.
|
|
11299
|
-
# @return [Integer]
|
|
11300
|
-
#
|
|
11301
|
-
# @see http://docs.aws.amazon.com/goto/WebAPI/qconnect-2020-10-19/TextAIPromptInferenceConfiguration AWS API Documentation
|
|
11302
|
-
#
|
|
11303
|
-
class TextAIPromptInferenceConfiguration < Struct.new(
|
|
11304
|
-
:temperature,
|
|
11305
|
-
:top_p,
|
|
11306
|
-
:top_k,
|
|
11307
|
-
:max_tokens_to_sample)
|
|
11308
|
-
SENSITIVE = [:temperature, :top_p, :top_k, :max_tokens_to_sample]
|
|
11309
|
-
include Aws::Structure
|
|
11310
|
-
end
|
|
11311
|
-
|
|
11312
11289
|
# Details about the source content text data.
|
|
11313
11290
|
#
|
|
11314
11291
|
# @!attribute [rw] title
|
data/lib/aws-sdk-qconnect.rb
CHANGED
data/sig/client.rbs
CHANGED
|
@@ -514,12 +514,10 @@ module Aws
|
|
|
514
514
|
?tags: Hash[::String, ::String],
|
|
515
515
|
?description: ::String,
|
|
516
516
|
?inference_configuration: {
|
|
517
|
-
|
|
518
|
-
|
|
519
|
-
|
|
520
|
-
|
|
521
|
-
max_tokens_to_sample: ::Integer?
|
|
522
|
-
}?
|
|
517
|
+
temperature: ::Float?,
|
|
518
|
+
top_p: ::Float?,
|
|
519
|
+
top_k: ::Integer?,
|
|
520
|
+
max_tokens_to_sample: ::Integer?
|
|
523
521
|
}
|
|
524
522
|
) -> _CreateAIPromptResponseSuccess
|
|
525
523
|
| (Hash[Symbol, untyped] params, ?Hash[Symbol, untyped] options) -> _CreateAIPromptResponseSuccess
|
|
@@ -2453,12 +2451,10 @@ module Aws
|
|
|
2453
2451
|
?description: ::String,
|
|
2454
2452
|
?model_id: ::String,
|
|
2455
2453
|
?inference_configuration: {
|
|
2456
|
-
|
|
2457
|
-
|
|
2458
|
-
|
|
2459
|
-
|
|
2460
|
-
max_tokens_to_sample: ::Integer?
|
|
2461
|
-
}?
|
|
2454
|
+
temperature: ::Float?,
|
|
2455
|
+
top_p: ::Float?,
|
|
2456
|
+
top_k: ::Integer?,
|
|
2457
|
+
max_tokens_to_sample: ::Integer?
|
|
2462
2458
|
}
|
|
2463
2459
|
) -> _UpdateAIPromptResponseSuccess
|
|
2464
2460
|
| (Hash[Symbol, untyped] params, ?Hash[Symbol, untyped] options) -> _UpdateAIPromptResponseSuccess
|
data/sig/types.rbs
CHANGED
|
@@ -183,14 +183,11 @@ module Aws::QConnect
|
|
|
183
183
|
end
|
|
184
184
|
|
|
185
185
|
class AIPromptInferenceConfiguration
|
|
186
|
-
attr_accessor
|
|
187
|
-
attr_accessor
|
|
188
|
-
|
|
189
|
-
|
|
190
|
-
|
|
191
|
-
end
|
|
192
|
-
class Unknown < AIPromptInferenceConfiguration
|
|
193
|
-
end
|
|
186
|
+
attr_accessor temperature: ::Float
|
|
187
|
+
attr_accessor top_p: ::Float
|
|
188
|
+
attr_accessor top_k: ::Integer
|
|
189
|
+
attr_accessor max_tokens_to_sample: ::Integer
|
|
190
|
+
SENSITIVE: [:temperature, :top_p, :top_k, :max_tokens_to_sample]
|
|
194
191
|
end
|
|
195
192
|
|
|
196
193
|
class AIPromptSummary
|
|
@@ -2995,14 +2992,6 @@ module Aws::QConnect
|
|
|
2995
2992
|
class TagResourceResponse < Aws::EmptyStructure
|
|
2996
2993
|
end
|
|
2997
2994
|
|
|
2998
|
-
class TextAIPromptInferenceConfiguration
|
|
2999
|
-
attr_accessor temperature: ::Float
|
|
3000
|
-
attr_accessor top_p: ::Float
|
|
3001
|
-
attr_accessor top_k: ::Integer
|
|
3002
|
-
attr_accessor max_tokens_to_sample: ::Integer
|
|
3003
|
-
SENSITIVE: [:temperature, :top_p, :top_k, :max_tokens_to_sample]
|
|
3004
|
-
end
|
|
3005
|
-
|
|
3006
2995
|
class TextData
|
|
3007
2996
|
attr_accessor title: Types::DocumentText
|
|
3008
2997
|
attr_accessor excerpt: Types::DocumentText
|
metadata
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
|
2
2
|
name: aws-sdk-qconnect
|
|
3
3
|
version: !ruby/object:Gem::Version
|
|
4
|
-
version: 1.
|
|
4
|
+
version: 1.48.0
|
|
5
5
|
platform: ruby
|
|
6
6
|
authors:
|
|
7
7
|
- Amazon Web Services
|
|
@@ -18,7 +18,7 @@ dependencies:
|
|
|
18
18
|
version: '3'
|
|
19
19
|
- - ">="
|
|
20
20
|
- !ruby/object:Gem::Version
|
|
21
|
-
version: 3.241.
|
|
21
|
+
version: 3.241.4
|
|
22
22
|
type: :runtime
|
|
23
23
|
prerelease: false
|
|
24
24
|
version_requirements: !ruby/object:Gem::Requirement
|
|
@@ -28,7 +28,7 @@ dependencies:
|
|
|
28
28
|
version: '3'
|
|
29
29
|
- - ">="
|
|
30
30
|
- !ruby/object:Gem::Version
|
|
31
|
-
version: 3.241.
|
|
31
|
+
version: 3.241.4
|
|
32
32
|
- !ruby/object:Gem::Dependency
|
|
33
33
|
name: aws-sigv4
|
|
34
34
|
requirement: !ruby/object:Gem::Requirement
|