aws-sdk-qconnect 1.46.0 → 1.48.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: d0fda67c4ba170886c591bd05d85cdd3d7b6ad112db408c698c89151d16abd95
4
- data.tar.gz: '08abd3e8d1e1c14873b730f7a93a73007d2a09d926073ac39c2e487d0d61a82f'
3
+ metadata.gz: 737e4ee1ca02ba610fb9d4d7b29439b4092b957e2d7a080d8407bf359c690560
4
+ data.tar.gz: eacec788d0fb123ffe936f7cab221b91357a808013cd245ce816b8b94023c9cd
5
5
  SHA512:
6
- metadata.gz: 93f33db072122132c431e274310c05ce43e0d1a82a386125cb480147aca0b99d59b1495ab5dff77149710b3542e5d8fa17edc8ed3610ce1cda58795c3c7ce9e5
7
- data.tar.gz: a7fabc3b11115a03b816c3ba08db3f37e7ea3a79083795ff5e1fcc71c12db0f3bc0425bfe9fb826d48bfaf8a655602ee816d4d652af33270de4ac6acec4cd53e
6
+ metadata.gz: 165120e837d8ade4a3fbef6e60bb36aef12066703dba30b99adb31ed4b0a232b369d5239368ce6c9c84844e395c89e266e04589f653d0b83b4a233228934a0c5
7
+ data.tar.gz: 9202e10c27266a6d7fb6944dc2ed621b843c3d70590199db1f506ff6534bd2ede4508d5e85c4619cf07aa8ad3c4878e135f97b3ed4a5f1f6d1094bbb63cd2456
data/CHANGELOG.md CHANGED
@@ -1,6 +1,16 @@
1
1
  Unreleased Changes
2
2
  ------------------
3
3
 
4
+ 1.48.0 (2026-01-16)
5
+ ------------------
6
+
7
+ * Feature - Code Generated Changes, see `./build_tools` or `aws-sdk-core`'s CHANGELOG.md for details.
8
+
9
+ 1.47.0 (2026-01-15)
10
+ ------------------
11
+
12
+ * Feature - Fix inference configuration shapes for the CreateAIPrompt and UpdateAIPrompt APIs, Modify Text Length Limit for SendMessage API
13
+
4
14
  1.46.0 (2026-01-08)
5
15
  ------------------
6
16
 
data/VERSION CHANGED
@@ -1 +1 @@
1
- 1.46.0
1
+ 1.48.0
@@ -1574,12 +1574,10 @@ module Aws::QConnect
1574
1574
  # },
1575
1575
  # description: "Description",
1576
1576
  # inference_configuration: {
1577
- # text_ai_prompt_inference_configuration: {
1578
- # temperature: 1.0,
1579
- # top_p: 1.0,
1580
- # top_k: 1,
1581
- # max_tokens_to_sample: 1,
1582
- # },
1577
+ # temperature: 1.0,
1578
+ # top_p: 1.0,
1579
+ # top_k: 1,
1580
+ # max_tokens_to_sample: 1,
1583
1581
  # },
1584
1582
  # })
1585
1583
  #
@@ -1595,10 +1593,10 @@ module Aws::QConnect
1595
1593
  # resp.ai_prompt.model_id #=> String
1596
1594
  # resp.ai_prompt.api_format #=> String, one of "ANTHROPIC_CLAUDE_MESSAGES", "ANTHROPIC_CLAUDE_TEXT_COMPLETIONS", "MESSAGES", "TEXT_COMPLETIONS"
1597
1595
  # resp.ai_prompt.template_configuration.text_full_ai_prompt_edit_template_configuration.text #=> String
1598
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.temperature #=> Float
1599
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.top_p #=> Float
1600
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.top_k #=> Integer
1601
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.max_tokens_to_sample #=> Integer
1596
+ # resp.ai_prompt.inference_configuration.temperature #=> Float
1597
+ # resp.ai_prompt.inference_configuration.top_p #=> Float
1598
+ # resp.ai_prompt.inference_configuration.top_k #=> Integer
1599
+ # resp.ai_prompt.inference_configuration.max_tokens_to_sample #=> Integer
1602
1600
  # resp.ai_prompt.modified_time #=> Time
1603
1601
  # resp.ai_prompt.description #=> String
1604
1602
  # resp.ai_prompt.visibility_status #=> String, one of "SAVED", "PUBLISHED"
@@ -1667,10 +1665,10 @@ module Aws::QConnect
1667
1665
  # resp.ai_prompt.model_id #=> String
1668
1666
  # resp.ai_prompt.api_format #=> String, one of "ANTHROPIC_CLAUDE_MESSAGES", "ANTHROPIC_CLAUDE_TEXT_COMPLETIONS", "MESSAGES", "TEXT_COMPLETIONS"
1669
1667
  # resp.ai_prompt.template_configuration.text_full_ai_prompt_edit_template_configuration.text #=> String
1670
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.temperature #=> Float
1671
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.top_p #=> Float
1672
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.top_k #=> Integer
1673
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.max_tokens_to_sample #=> Integer
1668
+ # resp.ai_prompt.inference_configuration.temperature #=> Float
1669
+ # resp.ai_prompt.inference_configuration.top_p #=> Float
1670
+ # resp.ai_prompt.inference_configuration.top_k #=> Integer
1671
+ # resp.ai_prompt.inference_configuration.max_tokens_to_sample #=> Integer
1674
1672
  # resp.ai_prompt.modified_time #=> Time
1675
1673
  # resp.ai_prompt.description #=> String
1676
1674
  # resp.ai_prompt.visibility_status #=> String, one of "SAVED", "PUBLISHED"
@@ -3992,10 +3990,10 @@ module Aws::QConnect
3992
3990
  # resp.ai_prompt.model_id #=> String
3993
3991
  # resp.ai_prompt.api_format #=> String, one of "ANTHROPIC_CLAUDE_MESSAGES", "ANTHROPIC_CLAUDE_TEXT_COMPLETIONS", "MESSAGES", "TEXT_COMPLETIONS"
3994
3992
  # resp.ai_prompt.template_configuration.text_full_ai_prompt_edit_template_configuration.text #=> String
3995
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.temperature #=> Float
3996
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.top_p #=> Float
3997
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.top_k #=> Integer
3998
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.max_tokens_to_sample #=> Integer
3993
+ # resp.ai_prompt.inference_configuration.temperature #=> Float
3994
+ # resp.ai_prompt.inference_configuration.top_p #=> Float
3995
+ # resp.ai_prompt.inference_configuration.top_k #=> Integer
3996
+ # resp.ai_prompt.inference_configuration.max_tokens_to_sample #=> Integer
3999
3997
  # resp.ai_prompt.modified_time #=> Time
4000
3998
  # resp.ai_prompt.description #=> String
4001
3999
  # resp.ai_prompt.visibility_status #=> String, one of "SAVED", "PUBLISHED"
@@ -8317,12 +8315,10 @@ module Aws::QConnect
8317
8315
  # description: "Description",
8318
8316
  # model_id: "AIPromptModelIdentifier",
8319
8317
  # inference_configuration: {
8320
- # text_ai_prompt_inference_configuration: {
8321
- # temperature: 1.0,
8322
- # top_p: 1.0,
8323
- # top_k: 1,
8324
- # max_tokens_to_sample: 1,
8325
- # },
8318
+ # temperature: 1.0,
8319
+ # top_p: 1.0,
8320
+ # top_k: 1,
8321
+ # max_tokens_to_sample: 1,
8326
8322
  # },
8327
8323
  # })
8328
8324
  #
@@ -8338,10 +8334,10 @@ module Aws::QConnect
8338
8334
  # resp.ai_prompt.model_id #=> String
8339
8335
  # resp.ai_prompt.api_format #=> String, one of "ANTHROPIC_CLAUDE_MESSAGES", "ANTHROPIC_CLAUDE_TEXT_COMPLETIONS", "MESSAGES", "TEXT_COMPLETIONS"
8340
8336
  # resp.ai_prompt.template_configuration.text_full_ai_prompt_edit_template_configuration.text #=> String
8341
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.temperature #=> Float
8342
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.top_p #=> Float
8343
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.top_k #=> Integer
8344
- # resp.ai_prompt.inference_configuration.text_ai_prompt_inference_configuration.max_tokens_to_sample #=> Integer
8337
+ # resp.ai_prompt.inference_configuration.temperature #=> Float
8338
+ # resp.ai_prompt.inference_configuration.top_p #=> Float
8339
+ # resp.ai_prompt.inference_configuration.top_k #=> Integer
8340
+ # resp.ai_prompt.inference_configuration.max_tokens_to_sample #=> Integer
8345
8341
  # resp.ai_prompt.modified_time #=> Time
8346
8342
  # resp.ai_prompt.description #=> String
8347
8343
  # resp.ai_prompt.visibility_status #=> String, one of "SAVED", "PUBLISHED"
@@ -9463,7 +9459,7 @@ module Aws::QConnect
9463
9459
  tracer: tracer
9464
9460
  )
9465
9461
  context[:gem_name] = 'aws-sdk-qconnect'
9466
- context[:gem_version] = '1.46.0'
9462
+ context[:gem_version] = '1.48.0'
9467
9463
  Seahorse::Client::Request.new(handlers, context)
9468
9464
  end
9469
9465
 
@@ -39,7 +39,7 @@ module Aws::QConnect
39
39
  AIGuardrailWordPolicyConfig = Shapes::StructureShape.new(name: 'AIGuardrailWordPolicyConfig')
40
40
  AIPromptAPIFormat = Shapes::StringShape.new(name: 'AIPromptAPIFormat')
41
41
  AIPromptData = Shapes::StructureShape.new(name: 'AIPromptData')
42
- AIPromptInferenceConfiguration = Shapes::UnionShape.new(name: 'AIPromptInferenceConfiguration')
42
+ AIPromptInferenceConfiguration = Shapes::StructureShape.new(name: 'AIPromptInferenceConfiguration')
43
43
  AIPromptModelIdentifier = Shapes::StringShape.new(name: 'AIPromptModelIdentifier')
44
44
  AIPromptSummary = Shapes::StructureShape.new(name: 'AIPromptSummary')
45
45
  AIPromptSummaryList = Shapes::ListShape.new(name: 'AIPromptSummaryList')
@@ -587,7 +587,6 @@ module Aws::QConnect
587
587
  Tags = Shapes::MapShape.new(name: 'Tags')
588
588
  TargetType = Shapes::StringShape.new(name: 'TargetType')
589
589
  TextAIPrompt = Shapes::StringShape.new(name: 'TextAIPrompt')
590
- TextAIPromptInferenceConfiguration = Shapes::StructureShape.new(name: 'TextAIPromptInferenceConfiguration')
591
590
  TextData = Shapes::StructureShape.new(name: 'TextData')
592
591
  TextFullAIPromptEditTemplateConfiguration = Shapes::StructureShape.new(name: 'TextFullAIPromptEditTemplateConfiguration')
593
592
  TextMessage = Shapes::StructureShape.new(name: 'TextMessage')
@@ -814,10 +813,10 @@ module Aws::QConnect
814
813
  AIPromptData.add_member(:status, Shapes::ShapeRef.new(shape: Status, location_name: "status"))
815
814
  AIPromptData.struct_class = Types::AIPromptData
816
815
 
817
- AIPromptInferenceConfiguration.add_member(:text_ai_prompt_inference_configuration, Shapes::ShapeRef.new(shape: TextAIPromptInferenceConfiguration, location_name: "textAIPromptInferenceConfiguration"))
818
- AIPromptInferenceConfiguration.add_member(:unknown, Shapes::ShapeRef.new(shape: nil, location_name: 'unknown'))
819
- AIPromptInferenceConfiguration.add_member_subclass(:text_ai_prompt_inference_configuration, Types::AIPromptInferenceConfiguration::TextAiPromptInferenceConfiguration)
820
- AIPromptInferenceConfiguration.add_member_subclass(:unknown, Types::AIPromptInferenceConfiguration::Unknown)
816
+ AIPromptInferenceConfiguration.add_member(:temperature, Shapes::ShapeRef.new(shape: Probability, location_name: "temperature"))
817
+ AIPromptInferenceConfiguration.add_member(:top_p, Shapes::ShapeRef.new(shape: Probability, location_name: "topP"))
818
+ AIPromptInferenceConfiguration.add_member(:top_k, Shapes::ShapeRef.new(shape: TopK, location_name: "topK"))
819
+ AIPromptInferenceConfiguration.add_member(:max_tokens_to_sample, Shapes::ShapeRef.new(shape: MaxTokensToSample, location_name: "maxTokensToSample"))
821
820
  AIPromptInferenceConfiguration.struct_class = Types::AIPromptInferenceConfiguration
822
821
 
823
822
  AIPromptSummary.add_member(:name, Shapes::ShapeRef.new(shape: Name, required: true, location_name: "name"))
@@ -2998,12 +2997,6 @@ module Aws::QConnect
2998
2997
  Tags.key = Shapes::ShapeRef.new(shape: TagKey)
2999
2998
  Tags.value = Shapes::ShapeRef.new(shape: TagValue)
3000
2999
 
3001
- TextAIPromptInferenceConfiguration.add_member(:temperature, Shapes::ShapeRef.new(shape: Probability, location_name: "temperature"))
3002
- TextAIPromptInferenceConfiguration.add_member(:top_p, Shapes::ShapeRef.new(shape: Probability, location_name: "topP"))
3003
- TextAIPromptInferenceConfiguration.add_member(:top_k, Shapes::ShapeRef.new(shape: TopK, location_name: "topK"))
3004
- TextAIPromptInferenceConfiguration.add_member(:max_tokens_to_sample, Shapes::ShapeRef.new(shape: MaxTokensToSample, location_name: "maxTokensToSample"))
3005
- TextAIPromptInferenceConfiguration.struct_class = Types::TextAIPromptInferenceConfiguration
3006
-
3007
3000
  TextData.add_member(:title, Shapes::ShapeRef.new(shape: DocumentText, location_name: "title"))
3008
3001
  TextData.add_member(:excerpt, Shapes::ShapeRef.new(shape: DocumentText, location_name: "excerpt"))
3009
3002
  TextData.struct_class = Types::TextData
@@ -692,25 +692,32 @@ module Aws::QConnect
692
692
 
693
693
  # The configuration for inference parameters when using AI Prompts.
694
694
  #
695
- # @note AIPromptInferenceConfiguration is a union - when making an API calls you must set exactly one of the members.
695
+ # @!attribute [rw] temperature
696
+ # The temperature setting for controlling randomness in the generated
697
+ # response.
698
+ # @return [Float]
696
699
  #
697
- # @note AIPromptInferenceConfiguration is a union - when returned from an API call exactly one value will be set and the returned type will be a subclass of AIPromptInferenceConfiguration corresponding to the set member.
700
+ # @!attribute [rw] top_p
701
+ # The top-P sampling parameter for nucleus sampling.
702
+ # @return [Float]
698
703
  #
699
- # @!attribute [rw] text_ai_prompt_inference_configuration
700
- # The inference configuration for text-based AI Prompts.
701
- # @return [Types::TextAIPromptInferenceConfiguration]
704
+ # @!attribute [rw] top_k
705
+ # The top-K sampling parameter for token selection.
706
+ # @return [Integer]
707
+ #
708
+ # @!attribute [rw] max_tokens_to_sample
709
+ # The maximum number of tokens to generate in the response.
710
+ # @return [Integer]
702
711
  #
703
712
  # @see http://docs.aws.amazon.com/goto/WebAPI/qconnect-2020-10-19/AIPromptInferenceConfiguration AWS API Documentation
704
713
  #
705
714
  class AIPromptInferenceConfiguration < Struct.new(
706
- :text_ai_prompt_inference_configuration,
707
- :unknown)
708
- SENSITIVE = []
715
+ :temperature,
716
+ :top_p,
717
+ :top_k,
718
+ :max_tokens_to_sample)
719
+ SENSITIVE = [:temperature, :top_p, :top_k, :max_tokens_to_sample]
709
720
  include Aws::Structure
710
- include Aws::Structure::Union
711
-
712
- class TextAiPromptInferenceConfiguration < AIPromptInferenceConfiguration; end
713
- class Unknown < AIPromptInferenceConfiguration; end
714
721
  end
715
722
 
716
723
  # The summary of the AI Prompt.
@@ -11279,36 +11286,6 @@ module Aws::QConnect
11279
11286
  #
11280
11287
  class TagResourceResponse < Aws::EmptyStructure; end
11281
11288
 
11282
- # Inference configuration for text-based AI Prompts.
11283
- #
11284
- # @!attribute [rw] temperature
11285
- # The temperature setting for controlling randomness in the generated
11286
- # response.
11287
- # @return [Float]
11288
- #
11289
- # @!attribute [rw] top_p
11290
- # The top-P sampling parameter for nucleus sampling.
11291
- # @return [Float]
11292
- #
11293
- # @!attribute [rw] top_k
11294
- # The top-K sampling parameter for token selection.
11295
- # @return [Integer]
11296
- #
11297
- # @!attribute [rw] max_tokens_to_sample
11298
- # The maximum number of tokens to generate in the response.
11299
- # @return [Integer]
11300
- #
11301
- # @see http://docs.aws.amazon.com/goto/WebAPI/qconnect-2020-10-19/TextAIPromptInferenceConfiguration AWS API Documentation
11302
- #
11303
- class TextAIPromptInferenceConfiguration < Struct.new(
11304
- :temperature,
11305
- :top_p,
11306
- :top_k,
11307
- :max_tokens_to_sample)
11308
- SENSITIVE = [:temperature, :top_p, :top_k, :max_tokens_to_sample]
11309
- include Aws::Structure
11310
- end
11311
-
11312
11289
  # Details about the source content text data.
11313
11290
  #
11314
11291
  # @!attribute [rw] title
@@ -55,7 +55,7 @@ module Aws::QConnect
55
55
  autoload :EndpointProvider, 'aws-sdk-qconnect/endpoint_provider'
56
56
  autoload :Endpoints, 'aws-sdk-qconnect/endpoints'
57
57
 
58
- GEM_VERSION = '1.46.0'
58
+ GEM_VERSION = '1.48.0'
59
59
 
60
60
  end
61
61
 
data/sig/client.rbs CHANGED
@@ -514,12 +514,10 @@ module Aws
514
514
  ?tags: Hash[::String, ::String],
515
515
  ?description: ::String,
516
516
  ?inference_configuration: {
517
- text_ai_prompt_inference_configuration: {
518
- temperature: ::Float?,
519
- top_p: ::Float?,
520
- top_k: ::Integer?,
521
- max_tokens_to_sample: ::Integer?
522
- }?
517
+ temperature: ::Float?,
518
+ top_p: ::Float?,
519
+ top_k: ::Integer?,
520
+ max_tokens_to_sample: ::Integer?
523
521
  }
524
522
  ) -> _CreateAIPromptResponseSuccess
525
523
  | (Hash[Symbol, untyped] params, ?Hash[Symbol, untyped] options) -> _CreateAIPromptResponseSuccess
@@ -2453,12 +2451,10 @@ module Aws
2453
2451
  ?description: ::String,
2454
2452
  ?model_id: ::String,
2455
2453
  ?inference_configuration: {
2456
- text_ai_prompt_inference_configuration: {
2457
- temperature: ::Float?,
2458
- top_p: ::Float?,
2459
- top_k: ::Integer?,
2460
- max_tokens_to_sample: ::Integer?
2461
- }?
2454
+ temperature: ::Float?,
2455
+ top_p: ::Float?,
2456
+ top_k: ::Integer?,
2457
+ max_tokens_to_sample: ::Integer?
2462
2458
  }
2463
2459
  ) -> _UpdateAIPromptResponseSuccess
2464
2460
  | (Hash[Symbol, untyped] params, ?Hash[Symbol, untyped] options) -> _UpdateAIPromptResponseSuccess
data/sig/types.rbs CHANGED
@@ -183,14 +183,11 @@ module Aws::QConnect
183
183
  end
184
184
 
185
185
  class AIPromptInferenceConfiguration
186
- attr_accessor text_ai_prompt_inference_configuration: Types::TextAIPromptInferenceConfiguration
187
- attr_accessor unknown: untyped
188
- SENSITIVE: []
189
-
190
- class TextAiPromptInferenceConfiguration < AIPromptInferenceConfiguration
191
- end
192
- class Unknown < AIPromptInferenceConfiguration
193
- end
186
+ attr_accessor temperature: ::Float
187
+ attr_accessor top_p: ::Float
188
+ attr_accessor top_k: ::Integer
189
+ attr_accessor max_tokens_to_sample: ::Integer
190
+ SENSITIVE: [:temperature, :top_p, :top_k, :max_tokens_to_sample]
194
191
  end
195
192
 
196
193
  class AIPromptSummary
@@ -2995,14 +2992,6 @@ module Aws::QConnect
2995
2992
  class TagResourceResponse < Aws::EmptyStructure
2996
2993
  end
2997
2994
 
2998
- class TextAIPromptInferenceConfiguration
2999
- attr_accessor temperature: ::Float
3000
- attr_accessor top_p: ::Float
3001
- attr_accessor top_k: ::Integer
3002
- attr_accessor max_tokens_to_sample: ::Integer
3003
- SENSITIVE: [:temperature, :top_p, :top_k, :max_tokens_to_sample]
3004
- end
3005
-
3006
2995
  class TextData
3007
2996
  attr_accessor title: Types::DocumentText
3008
2997
  attr_accessor excerpt: Types::DocumentText
metadata CHANGED
@@ -1,7 +1,7 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: aws-sdk-qconnect
3
3
  version: !ruby/object:Gem::Version
4
- version: 1.46.0
4
+ version: 1.48.0
5
5
  platform: ruby
6
6
  authors:
7
7
  - Amazon Web Services
@@ -18,7 +18,7 @@ dependencies:
18
18
  version: '3'
19
19
  - - ">="
20
20
  - !ruby/object:Gem::Version
21
- version: 3.241.3
21
+ version: 3.241.4
22
22
  type: :runtime
23
23
  prerelease: false
24
24
  version_requirements: !ruby/object:Gem::Requirement
@@ -28,7 +28,7 @@ dependencies:
28
28
  version: '3'
29
29
  - - ">="
30
30
  - !ruby/object:Gem::Version
31
- version: 3.241.3
31
+ version: 3.241.4
32
32
  - !ruby/object:Gem::Dependency
33
33
  name: aws-sigv4
34
34
  requirement: !ruby/object:Gem::Requirement