alibabacloud-aimiaobi20230801 1.30.0__tar.gz → 1.30.2__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/ChangeLog.md +12 -0
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/PKG-INFO +1 -1
- alibabacloud_aimiaobi20230801-1.30.2/alibabacloud_aimiaobi20230801/__init__.py +1 -0
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/alibabacloud_aimiaobi20230801/client.py +44 -4
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/alibabacloud_aimiaobi20230801/models.py +84 -0
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/alibabacloud_aimiaobi20230801.egg-info/PKG-INFO +1 -1
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/setup.py +1 -1
- alibabacloud_aimiaobi20230801-1.30.0/alibabacloud_aimiaobi20230801/__init__.py +0 -1
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/LICENSE +0 -0
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/MANIFEST.in +0 -0
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/README-CN.md +0 -0
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/README.md +0 -0
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/alibabacloud_aimiaobi20230801.egg-info/SOURCES.txt +0 -0
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/alibabacloud_aimiaobi20230801.egg-info/dependency_links.txt +0 -0
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/alibabacloud_aimiaobi20230801.egg-info/requires.txt +0 -0
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/alibabacloud_aimiaobi20230801.egg-info/top_level.txt +0 -0
- {alibabacloud_aimiaobi20230801-1.30.0 → alibabacloud_aimiaobi20230801-1.30.2}/setup.cfg +0 -0
|
@@ -1,3 +1,15 @@
|
|
|
1
|
+
2025-06-23 Version: 1.30.1
|
|
2
|
+
- Update API RunSearchGeneration: add request parameters ChatConfig.ExcludeGenerateOptions.
|
|
3
|
+
|
|
4
|
+
|
|
5
|
+
2025-06-18 Version: 1.30.0
|
|
6
|
+
- Support API AsyncCreateClipsTask.
|
|
7
|
+
- Support API AsyncCreateClipsTimeLine.
|
|
8
|
+
- Support API AsyncEditTimeline.
|
|
9
|
+
- Support API AsyncUploadVideo.
|
|
10
|
+
- Support API GetAutoClipsTaskInfo.
|
|
11
|
+
|
|
12
|
+
|
|
1
13
|
2025-06-16 Version: 1.29.0
|
|
2
14
|
- Support API ExportAuditContentResult.
|
|
3
15
|
- Support API GetSmartAuditResult.
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
__version__ = '1.30.2'
|
|
@@ -163,7 +163,7 @@ class Client(OpenApiClient):
|
|
|
163
163
|
runtime: util_models.RuntimeOptions,
|
|
164
164
|
) -> ai_miao_bi_20230801_models.AsyncCreateClipsTaskResponse:
|
|
165
165
|
"""
|
|
166
|
-
@summary
|
|
166
|
+
@summary 生成剪辑视频任务
|
|
167
167
|
|
|
168
168
|
@param tmp_req: AsyncCreateClipsTaskRequest
|
|
169
169
|
@param runtime: runtime options for this request RuntimeOptions
|
|
@@ -220,7 +220,7 @@ class Client(OpenApiClient):
|
|
|
220
220
|
runtime: util_models.RuntimeOptions,
|
|
221
221
|
) -> ai_miao_bi_20230801_models.AsyncCreateClipsTaskResponse:
|
|
222
222
|
"""
|
|
223
|
-
@summary
|
|
223
|
+
@summary 生成剪辑视频任务
|
|
224
224
|
|
|
225
225
|
@param tmp_req: AsyncCreateClipsTaskRequest
|
|
226
226
|
@param runtime: runtime options for this request RuntimeOptions
|
|
@@ -276,7 +276,7 @@ class Client(OpenApiClient):
|
|
|
276
276
|
request: ai_miao_bi_20230801_models.AsyncCreateClipsTaskRequest,
|
|
277
277
|
) -> ai_miao_bi_20230801_models.AsyncCreateClipsTaskResponse:
|
|
278
278
|
"""
|
|
279
|
-
@summary
|
|
279
|
+
@summary 生成剪辑视频任务
|
|
280
280
|
|
|
281
281
|
@param request: AsyncCreateClipsTaskRequest
|
|
282
282
|
@return: AsyncCreateClipsTaskResponse
|
|
@@ -289,7 +289,7 @@ class Client(OpenApiClient):
|
|
|
289
289
|
request: ai_miao_bi_20230801_models.AsyncCreateClipsTaskRequest,
|
|
290
290
|
) -> ai_miao_bi_20230801_models.AsyncCreateClipsTaskResponse:
|
|
291
291
|
"""
|
|
292
|
-
@summary
|
|
292
|
+
@summary 生成剪辑视频任务
|
|
293
293
|
|
|
294
294
|
@param request: AsyncCreateClipsTaskRequest
|
|
295
295
|
@return: AsyncCreateClipsTaskResponse
|
|
@@ -11791,6 +11791,8 @@ class Client(OpenApiClient):
|
|
|
11791
11791
|
body['Length'] = request.length
|
|
11792
11792
|
if not UtilClient.is_unset(request.length_range_shrink):
|
|
11793
11793
|
body['LengthRange'] = request.length_range_shrink
|
|
11794
|
+
if not UtilClient.is_unset(request.model_id):
|
|
11795
|
+
body['ModelId'] = request.model_id
|
|
11794
11796
|
if not UtilClient.is_unset(request.num_comments):
|
|
11795
11797
|
body['NumComments'] = request.num_comments
|
|
11796
11798
|
if not UtilClient.is_unset(request.sentiment_shrink):
|
|
@@ -11854,6 +11856,8 @@ class Client(OpenApiClient):
|
|
|
11854
11856
|
body['Length'] = request.length
|
|
11855
11857
|
if not UtilClient.is_unset(request.length_range_shrink):
|
|
11856
11858
|
body['LengthRange'] = request.length_range_shrink
|
|
11859
|
+
if not UtilClient.is_unset(request.model_id):
|
|
11860
|
+
body['ModelId'] = request.model_id
|
|
11857
11861
|
if not UtilClient.is_unset(request.num_comments):
|
|
11858
11862
|
body['NumComments'] = request.num_comments
|
|
11859
11863
|
if not UtilClient.is_unset(request.sentiment_shrink):
|
|
@@ -12283,6 +12287,8 @@ class Client(OpenApiClient):
|
|
|
12283
12287
|
body['CleanCache'] = request.clean_cache
|
|
12284
12288
|
if not UtilClient.is_unset(request.doc_id):
|
|
12285
12289
|
body['DocId'] = request.doc_id
|
|
12290
|
+
if not UtilClient.is_unset(request.model_name):
|
|
12291
|
+
body['ModelName'] = request.model_name
|
|
12286
12292
|
if not UtilClient.is_unset(request.node_number):
|
|
12287
12293
|
body['NodeNumber'] = request.node_number
|
|
12288
12294
|
if not UtilClient.is_unset(request.prompt):
|
|
@@ -12332,6 +12338,8 @@ class Client(OpenApiClient):
|
|
|
12332
12338
|
body['CleanCache'] = request.clean_cache
|
|
12333
12339
|
if not UtilClient.is_unset(request.doc_id):
|
|
12334
12340
|
body['DocId'] = request.doc_id
|
|
12341
|
+
if not UtilClient.is_unset(request.model_name):
|
|
12342
|
+
body['ModelName'] = request.model_name
|
|
12335
12343
|
if not UtilClient.is_unset(request.node_number):
|
|
12336
12344
|
body['NodeNumber'] = request.node_number
|
|
12337
12345
|
if not UtilClient.is_unset(request.prompt):
|
|
@@ -12411,6 +12419,8 @@ class Client(OpenApiClient):
|
|
|
12411
12419
|
body['IntroductionPrompt'] = request.introduction_prompt
|
|
12412
12420
|
if not UtilClient.is_unset(request.key_point_prompt):
|
|
12413
12421
|
body['KeyPointPrompt'] = request.key_point_prompt
|
|
12422
|
+
if not UtilClient.is_unset(request.model_name):
|
|
12423
|
+
body['ModelName'] = request.model_name
|
|
12414
12424
|
if not UtilClient.is_unset(request.session_id):
|
|
12415
12425
|
body['SessionId'] = request.session_id
|
|
12416
12426
|
if not UtilClient.is_unset(request.summary_prompt):
|
|
@@ -12460,6 +12470,8 @@ class Client(OpenApiClient):
|
|
|
12460
12470
|
body['IntroductionPrompt'] = request.introduction_prompt
|
|
12461
12471
|
if not UtilClient.is_unset(request.key_point_prompt):
|
|
12462
12472
|
body['KeyPointPrompt'] = request.key_point_prompt
|
|
12473
|
+
if not UtilClient.is_unset(request.model_name):
|
|
12474
|
+
body['ModelName'] = request.model_name
|
|
12463
12475
|
if not UtilClient.is_unset(request.session_id):
|
|
12464
12476
|
body['SessionId'] = request.session_id
|
|
12465
12477
|
if not UtilClient.is_unset(request.summary_prompt):
|
|
@@ -12541,6 +12553,8 @@ class Client(OpenApiClient):
|
|
|
12541
12553
|
body['ConversationContexts'] = request.conversation_contexts_shrink
|
|
12542
12554
|
if not UtilClient.is_unset(request.doc_ids_shrink):
|
|
12543
12555
|
body['DocIds'] = request.doc_ids_shrink
|
|
12556
|
+
if not UtilClient.is_unset(request.model_name):
|
|
12557
|
+
body['ModelName'] = request.model_name
|
|
12544
12558
|
if not UtilClient.is_unset(request.query):
|
|
12545
12559
|
body['Query'] = request.query
|
|
12546
12560
|
if not UtilClient.is_unset(request.reference_content):
|
|
@@ -12598,6 +12612,8 @@ class Client(OpenApiClient):
|
|
|
12598
12612
|
body['ConversationContexts'] = request.conversation_contexts_shrink
|
|
12599
12613
|
if not UtilClient.is_unset(request.doc_ids_shrink):
|
|
12600
12614
|
body['DocIds'] = request.doc_ids_shrink
|
|
12615
|
+
if not UtilClient.is_unset(request.model_name):
|
|
12616
|
+
body['ModelName'] = request.model_name
|
|
12601
12617
|
if not UtilClient.is_unset(request.query):
|
|
12602
12618
|
body['Query'] = request.query
|
|
12603
12619
|
if not UtilClient.is_unset(request.reference_content):
|
|
@@ -12669,6 +12685,8 @@ class Client(OpenApiClient):
|
|
|
12669
12685
|
body = {}
|
|
12670
12686
|
if not UtilClient.is_unset(request.doc_id):
|
|
12671
12687
|
body['DocId'] = request.doc_id
|
|
12688
|
+
if not UtilClient.is_unset(request.model_name):
|
|
12689
|
+
body['ModelName'] = request.model_name
|
|
12672
12690
|
if not UtilClient.is_unset(request.prompt):
|
|
12673
12691
|
body['Prompt'] = request.prompt
|
|
12674
12692
|
if not UtilClient.is_unset(request.session_id):
|
|
@@ -12710,6 +12728,8 @@ class Client(OpenApiClient):
|
|
|
12710
12728
|
body = {}
|
|
12711
12729
|
if not UtilClient.is_unset(request.doc_id):
|
|
12712
12730
|
body['DocId'] = request.doc_id
|
|
12731
|
+
if not UtilClient.is_unset(request.model_name):
|
|
12732
|
+
body['ModelName'] = request.model_name
|
|
12713
12733
|
if not UtilClient.is_unset(request.prompt):
|
|
12714
12734
|
body['Prompt'] = request.prompt
|
|
12715
12735
|
if not UtilClient.is_unset(request.session_id):
|
|
@@ -12899,6 +12919,8 @@ class Client(OpenApiClient):
|
|
|
12899
12919
|
body['CleanCache'] = request.clean_cache
|
|
12900
12920
|
if not UtilClient.is_unset(request.doc_id):
|
|
12901
12921
|
body['DocId'] = request.doc_id
|
|
12922
|
+
if not UtilClient.is_unset(request.model_name):
|
|
12923
|
+
body['ModelName'] = request.model_name
|
|
12902
12924
|
if not UtilClient.is_unset(request.recommend_content):
|
|
12903
12925
|
body['RecommendContent'] = request.recommend_content
|
|
12904
12926
|
if not UtilClient.is_unset(request.session_id):
|
|
@@ -12944,6 +12966,8 @@ class Client(OpenApiClient):
|
|
|
12944
12966
|
body['CleanCache'] = request.clean_cache
|
|
12945
12967
|
if not UtilClient.is_unset(request.doc_id):
|
|
12946
12968
|
body['DocId'] = request.doc_id
|
|
12969
|
+
if not UtilClient.is_unset(request.model_name):
|
|
12970
|
+
body['ModelName'] = request.model_name
|
|
12947
12971
|
if not UtilClient.is_unset(request.recommend_content):
|
|
12948
12972
|
body['RecommendContent'] = request.recommend_content
|
|
12949
12973
|
if not UtilClient.is_unset(request.session_id):
|
|
@@ -13011,6 +13035,8 @@ class Client(OpenApiClient):
|
|
|
13011
13035
|
"""
|
|
13012
13036
|
UtilClient.validate_model(request)
|
|
13013
13037
|
body = {}
|
|
13038
|
+
if not UtilClient.is_unset(request.model_id):
|
|
13039
|
+
body['ModelId'] = request.model_id
|
|
13014
13040
|
if not UtilClient.is_unset(request.prompt):
|
|
13015
13041
|
body['Prompt'] = request.prompt
|
|
13016
13042
|
if not UtilClient.is_unset(request.reference_content):
|
|
@@ -13060,6 +13086,8 @@ class Client(OpenApiClient):
|
|
|
13060
13086
|
"""
|
|
13061
13087
|
UtilClient.validate_model(request)
|
|
13062
13088
|
body = {}
|
|
13089
|
+
if not UtilClient.is_unset(request.model_id):
|
|
13090
|
+
body['ModelId'] = request.model_id
|
|
13063
13091
|
if not UtilClient.is_unset(request.prompt):
|
|
13064
13092
|
body['Prompt'] = request.prompt
|
|
13065
13093
|
if not UtilClient.is_unset(request.reference_content):
|
|
@@ -13237,6 +13265,8 @@ class Client(OpenApiClient):
|
|
|
13237
13265
|
body = {}
|
|
13238
13266
|
if not UtilClient.is_unset(request.doc_id):
|
|
13239
13267
|
body['DocId'] = request.doc_id
|
|
13268
|
+
if not UtilClient.is_unset(request.model_name):
|
|
13269
|
+
body['ModelName'] = request.model_name
|
|
13240
13270
|
if not UtilClient.is_unset(request.reference_content):
|
|
13241
13271
|
body['ReferenceContent'] = request.reference_content
|
|
13242
13272
|
if not UtilClient.is_unset(request.session_id):
|
|
@@ -13278,6 +13308,8 @@ class Client(OpenApiClient):
|
|
|
13278
13308
|
body = {}
|
|
13279
13309
|
if not UtilClient.is_unset(request.doc_id):
|
|
13280
13310
|
body['DocId'] = request.doc_id
|
|
13311
|
+
if not UtilClient.is_unset(request.model_name):
|
|
13312
|
+
body['ModelName'] = request.model_name
|
|
13281
13313
|
if not UtilClient.is_unset(request.reference_content):
|
|
13282
13314
|
body['ReferenceContent'] = request.reference_content
|
|
13283
13315
|
if not UtilClient.is_unset(request.session_id):
|
|
@@ -13345,6 +13377,8 @@ class Client(OpenApiClient):
|
|
|
13345
13377
|
body = {}
|
|
13346
13378
|
if not UtilClient.is_unset(request.doc_id):
|
|
13347
13379
|
body['DocId'] = request.doc_id
|
|
13380
|
+
if not UtilClient.is_unset(request.model_name):
|
|
13381
|
+
body['ModelName'] = request.model_name
|
|
13348
13382
|
if not UtilClient.is_unset(request.prompt):
|
|
13349
13383
|
body['Prompt'] = request.prompt
|
|
13350
13384
|
if not UtilClient.is_unset(request.reference_content):
|
|
@@ -13388,6 +13422,8 @@ class Client(OpenApiClient):
|
|
|
13388
13422
|
body = {}
|
|
13389
13423
|
if not UtilClient.is_unset(request.doc_id):
|
|
13390
13424
|
body['DocId'] = request.doc_id
|
|
13425
|
+
if not UtilClient.is_unset(request.model_name):
|
|
13426
|
+
body['ModelName'] = request.model_name
|
|
13391
13427
|
if not UtilClient.is_unset(request.prompt):
|
|
13392
13428
|
body['Prompt'] = request.prompt
|
|
13393
13429
|
if not UtilClient.is_unset(request.reference_content):
|
|
@@ -13575,6 +13611,8 @@ class Client(OpenApiClient):
|
|
|
13575
13611
|
body['DocIds'] = request.doc_ids_shrink
|
|
13576
13612
|
if not UtilClient.is_unset(request.key_point_prompt):
|
|
13577
13613
|
body['KeyPointPrompt'] = request.key_point_prompt
|
|
13614
|
+
if not UtilClient.is_unset(request.model_name):
|
|
13615
|
+
body['ModelName'] = request.model_name
|
|
13578
13616
|
if not UtilClient.is_unset(request.session_id):
|
|
13579
13617
|
body['SessionId'] = request.session_id
|
|
13580
13618
|
if not UtilClient.is_unset(request.summary_prompt):
|
|
@@ -13622,6 +13660,8 @@ class Client(OpenApiClient):
|
|
|
13622
13660
|
body['DocIds'] = request.doc_ids_shrink
|
|
13623
13661
|
if not UtilClient.is_unset(request.key_point_prompt):
|
|
13624
13662
|
body['KeyPointPrompt'] = request.key_point_prompt
|
|
13663
|
+
if not UtilClient.is_unset(request.model_name):
|
|
13664
|
+
body['ModelName'] = request.model_name
|
|
13625
13665
|
if not UtilClient.is_unset(request.session_id):
|
|
13626
13666
|
body['SessionId'] = request.session_id
|
|
13627
13667
|
if not UtilClient.is_unset(request.summary_prompt):
|
|
@@ -30542,6 +30542,7 @@ class RunCommentGenerationRequest(TeaModel):
|
|
|
30542
30542
|
extra_info: str = None,
|
|
30543
30543
|
length: str = None,
|
|
30544
30544
|
length_range: Dict[str, Any] = None,
|
|
30545
|
+
model_id: str = None,
|
|
30545
30546
|
num_comments: str = None,
|
|
30546
30547
|
sentiment: Dict[str, Any] = None,
|
|
30547
30548
|
session_id: str = None,
|
|
@@ -30555,6 +30556,7 @@ class RunCommentGenerationRequest(TeaModel):
|
|
|
30555
30556
|
self.length = length
|
|
30556
30557
|
# This parameter is required.
|
|
30557
30558
|
self.length_range = length_range
|
|
30559
|
+
self.model_id = model_id
|
|
30558
30560
|
# This parameter is required.
|
|
30559
30561
|
self.num_comments = num_comments
|
|
30560
30562
|
# This parameter is required.
|
|
@@ -30585,6 +30587,8 @@ class RunCommentGenerationRequest(TeaModel):
|
|
|
30585
30587
|
result['Length'] = self.length
|
|
30586
30588
|
if self.length_range is not None:
|
|
30587
30589
|
result['LengthRange'] = self.length_range
|
|
30590
|
+
if self.model_id is not None:
|
|
30591
|
+
result['ModelId'] = self.model_id
|
|
30588
30592
|
if self.num_comments is not None:
|
|
30589
30593
|
result['NumComments'] = self.num_comments
|
|
30590
30594
|
if self.sentiment is not None:
|
|
@@ -30611,6 +30615,8 @@ class RunCommentGenerationRequest(TeaModel):
|
|
|
30611
30615
|
self.length = m.get('Length')
|
|
30612
30616
|
if m.get('LengthRange') is not None:
|
|
30613
30617
|
self.length_range = m.get('LengthRange')
|
|
30618
|
+
if m.get('ModelId') is not None:
|
|
30619
|
+
self.model_id = m.get('ModelId')
|
|
30614
30620
|
if m.get('NumComments') is not None:
|
|
30615
30621
|
self.num_comments = m.get('NumComments')
|
|
30616
30622
|
if m.get('Sentiment') is not None:
|
|
@@ -30635,6 +30641,7 @@ class RunCommentGenerationShrinkRequest(TeaModel):
|
|
|
30635
30641
|
extra_info: str = None,
|
|
30636
30642
|
length: str = None,
|
|
30637
30643
|
length_range_shrink: str = None,
|
|
30644
|
+
model_id: str = None,
|
|
30638
30645
|
num_comments: str = None,
|
|
30639
30646
|
sentiment_shrink: str = None,
|
|
30640
30647
|
session_id: str = None,
|
|
@@ -30648,6 +30655,7 @@ class RunCommentGenerationShrinkRequest(TeaModel):
|
|
|
30648
30655
|
self.length = length
|
|
30649
30656
|
# This parameter is required.
|
|
30650
30657
|
self.length_range_shrink = length_range_shrink
|
|
30658
|
+
self.model_id = model_id
|
|
30651
30659
|
# This parameter is required.
|
|
30652
30660
|
self.num_comments = num_comments
|
|
30653
30661
|
# This parameter is required.
|
|
@@ -30678,6 +30686,8 @@ class RunCommentGenerationShrinkRequest(TeaModel):
|
|
|
30678
30686
|
result['Length'] = self.length
|
|
30679
30687
|
if self.length_range_shrink is not None:
|
|
30680
30688
|
result['LengthRange'] = self.length_range_shrink
|
|
30689
|
+
if self.model_id is not None:
|
|
30690
|
+
result['ModelId'] = self.model_id
|
|
30681
30691
|
if self.num_comments is not None:
|
|
30682
30692
|
result['NumComments'] = self.num_comments
|
|
30683
30693
|
if self.sentiment_shrink is not None:
|
|
@@ -30704,6 +30714,8 @@ class RunCommentGenerationShrinkRequest(TeaModel):
|
|
|
30704
30714
|
self.length = m.get('Length')
|
|
30705
30715
|
if m.get('LengthRange') is not None:
|
|
30706
30716
|
self.length_range_shrink = m.get('LengthRange')
|
|
30717
|
+
if m.get('ModelId') is not None:
|
|
30718
|
+
self.model_id = m.get('ModelId')
|
|
30707
30719
|
if m.get('NumComments') is not None:
|
|
30708
30720
|
self.num_comments = m.get('NumComments')
|
|
30709
30721
|
if m.get('Sentiment') is not None:
|
|
@@ -32156,6 +32168,7 @@ class RunDocBrainmapRequest(TeaModel):
|
|
|
32156
32168
|
self,
|
|
32157
32169
|
clean_cache: bool = None,
|
|
32158
32170
|
doc_id: str = None,
|
|
32171
|
+
model_name: str = None,
|
|
32159
32172
|
node_number: int = None,
|
|
32160
32173
|
prompt: str = None,
|
|
32161
32174
|
session_id: str = None,
|
|
@@ -32166,6 +32179,7 @@ class RunDocBrainmapRequest(TeaModel):
|
|
|
32166
32179
|
self.clean_cache = clean_cache
|
|
32167
32180
|
# This parameter is required.
|
|
32168
32181
|
self.doc_id = doc_id
|
|
32182
|
+
self.model_name = model_name
|
|
32169
32183
|
self.node_number = node_number
|
|
32170
32184
|
self.prompt = prompt
|
|
32171
32185
|
# This parameter is required.
|
|
@@ -32188,6 +32202,8 @@ class RunDocBrainmapRequest(TeaModel):
|
|
|
32188
32202
|
result['CleanCache'] = self.clean_cache
|
|
32189
32203
|
if self.doc_id is not None:
|
|
32190
32204
|
result['DocId'] = self.doc_id
|
|
32205
|
+
if self.model_name is not None:
|
|
32206
|
+
result['ModelName'] = self.model_name
|
|
32191
32207
|
if self.node_number is not None:
|
|
32192
32208
|
result['NodeNumber'] = self.node_number
|
|
32193
32209
|
if self.prompt is not None:
|
|
@@ -32208,6 +32224,8 @@ class RunDocBrainmapRequest(TeaModel):
|
|
|
32208
32224
|
self.clean_cache = m.get('CleanCache')
|
|
32209
32225
|
if m.get('DocId') is not None:
|
|
32210
32226
|
self.doc_id = m.get('DocId')
|
|
32227
|
+
if m.get('ModelName') is not None:
|
|
32228
|
+
self.model_name = m.get('ModelName')
|
|
32211
32229
|
if m.get('NodeNumber') is not None:
|
|
32212
32230
|
self.node_number = m.get('NodeNumber')
|
|
32213
32231
|
if m.get('Prompt') is not None:
|
|
@@ -32482,6 +32500,7 @@ class RunDocIntroductionRequest(TeaModel):
|
|
|
32482
32500
|
doc_id: str = None,
|
|
32483
32501
|
introduction_prompt: str = None,
|
|
32484
32502
|
key_point_prompt: str = None,
|
|
32503
|
+
model_name: str = None,
|
|
32485
32504
|
session_id: str = None,
|
|
32486
32505
|
summary_prompt: str = None,
|
|
32487
32506
|
workspace_id: str = None,
|
|
@@ -32492,6 +32511,7 @@ class RunDocIntroductionRequest(TeaModel):
|
|
|
32492
32511
|
self.doc_id = doc_id
|
|
32493
32512
|
self.introduction_prompt = introduction_prompt
|
|
32494
32513
|
self.key_point_prompt = key_point_prompt
|
|
32514
|
+
self.model_name = model_name
|
|
32495
32515
|
# This parameter is required.
|
|
32496
32516
|
self.session_id = session_id
|
|
32497
32517
|
self.summary_prompt = summary_prompt
|
|
@@ -32516,6 +32536,8 @@ class RunDocIntroductionRequest(TeaModel):
|
|
|
32516
32536
|
result['IntroductionPrompt'] = self.introduction_prompt
|
|
32517
32537
|
if self.key_point_prompt is not None:
|
|
32518
32538
|
result['KeyPointPrompt'] = self.key_point_prompt
|
|
32539
|
+
if self.model_name is not None:
|
|
32540
|
+
result['ModelName'] = self.model_name
|
|
32519
32541
|
if self.session_id is not None:
|
|
32520
32542
|
result['SessionId'] = self.session_id
|
|
32521
32543
|
if self.summary_prompt is not None:
|
|
@@ -32536,6 +32558,8 @@ class RunDocIntroductionRequest(TeaModel):
|
|
|
32536
32558
|
self.introduction_prompt = m.get('IntroductionPrompt')
|
|
32537
32559
|
if m.get('KeyPointPrompt') is not None:
|
|
32538
32560
|
self.key_point_prompt = m.get('KeyPointPrompt')
|
|
32561
|
+
if m.get('ModelName') is not None:
|
|
32562
|
+
self.model_name = m.get('ModelName')
|
|
32539
32563
|
if m.get('SessionId') is not None:
|
|
32540
32564
|
self.session_id = m.get('SessionId')
|
|
32541
32565
|
if m.get('SummaryPrompt') is not None:
|
|
@@ -32974,6 +32998,7 @@ class RunDocQaRequest(TeaModel):
|
|
|
32974
32998
|
category_ids: List[str] = None,
|
|
32975
32999
|
conversation_contexts: List[RunDocQaRequestConversationContexts] = None,
|
|
32976
33000
|
doc_ids: List[str] = None,
|
|
33001
|
+
model_name: str = None,
|
|
32977
33002
|
query: str = None,
|
|
32978
33003
|
reference_content: str = None,
|
|
32979
33004
|
search_source: str = None,
|
|
@@ -32983,6 +33008,7 @@ class RunDocQaRequest(TeaModel):
|
|
|
32983
33008
|
self.category_ids = category_ids
|
|
32984
33009
|
self.conversation_contexts = conversation_contexts
|
|
32985
33010
|
self.doc_ids = doc_ids
|
|
33011
|
+
self.model_name = model_name
|
|
32986
33012
|
# This parameter is required.
|
|
32987
33013
|
self.query = query
|
|
32988
33014
|
self.reference_content = reference_content
|
|
@@ -33013,6 +33039,8 @@ class RunDocQaRequest(TeaModel):
|
|
|
33013
33039
|
result['ConversationContexts'].append(k.to_map() if k else None)
|
|
33014
33040
|
if self.doc_ids is not None:
|
|
33015
33041
|
result['DocIds'] = self.doc_ids
|
|
33042
|
+
if self.model_name is not None:
|
|
33043
|
+
result['ModelName'] = self.model_name
|
|
33016
33044
|
if self.query is not None:
|
|
33017
33045
|
result['Query'] = self.query
|
|
33018
33046
|
if self.reference_content is not None:
|
|
@@ -33036,6 +33064,8 @@ class RunDocQaRequest(TeaModel):
|
|
|
33036
33064
|
self.conversation_contexts.append(temp_model.from_map(k))
|
|
33037
33065
|
if m.get('DocIds') is not None:
|
|
33038
33066
|
self.doc_ids = m.get('DocIds')
|
|
33067
|
+
if m.get('ModelName') is not None:
|
|
33068
|
+
self.model_name = m.get('ModelName')
|
|
33039
33069
|
if m.get('Query') is not None:
|
|
33040
33070
|
self.query = m.get('Query')
|
|
33041
33071
|
if m.get('ReferenceContent') is not None:
|
|
@@ -33055,6 +33085,7 @@ class RunDocQaShrinkRequest(TeaModel):
|
|
|
33055
33085
|
category_ids_shrink: str = None,
|
|
33056
33086
|
conversation_contexts_shrink: str = None,
|
|
33057
33087
|
doc_ids_shrink: str = None,
|
|
33088
|
+
model_name: str = None,
|
|
33058
33089
|
query: str = None,
|
|
33059
33090
|
reference_content: str = None,
|
|
33060
33091
|
search_source: str = None,
|
|
@@ -33064,6 +33095,7 @@ class RunDocQaShrinkRequest(TeaModel):
|
|
|
33064
33095
|
self.category_ids_shrink = category_ids_shrink
|
|
33065
33096
|
self.conversation_contexts_shrink = conversation_contexts_shrink
|
|
33066
33097
|
self.doc_ids_shrink = doc_ids_shrink
|
|
33098
|
+
self.model_name = model_name
|
|
33067
33099
|
# This parameter is required.
|
|
33068
33100
|
self.query = query
|
|
33069
33101
|
self.reference_content = reference_content
|
|
@@ -33089,6 +33121,8 @@ class RunDocQaShrinkRequest(TeaModel):
|
|
|
33089
33121
|
result['ConversationContexts'] = self.conversation_contexts_shrink
|
|
33090
33122
|
if self.doc_ids_shrink is not None:
|
|
33091
33123
|
result['DocIds'] = self.doc_ids_shrink
|
|
33124
|
+
if self.model_name is not None:
|
|
33125
|
+
result['ModelName'] = self.model_name
|
|
33092
33126
|
if self.query is not None:
|
|
33093
33127
|
result['Query'] = self.query
|
|
33094
33128
|
if self.reference_content is not None:
|
|
@@ -33109,6 +33143,8 @@ class RunDocQaShrinkRequest(TeaModel):
|
|
|
33109
33143
|
self.conversation_contexts_shrink = m.get('ConversationContexts')
|
|
33110
33144
|
if m.get('DocIds') is not None:
|
|
33111
33145
|
self.doc_ids_shrink = m.get('DocIds')
|
|
33146
|
+
if m.get('ModelName') is not None:
|
|
33147
|
+
self.model_name = m.get('ModelName')
|
|
33112
33148
|
if m.get('Query') is not None:
|
|
33113
33149
|
self.query = m.get('Query')
|
|
33114
33150
|
if m.get('ReferenceContent') is not None:
|
|
@@ -33598,12 +33634,14 @@ class RunDocSmartCardRequest(TeaModel):
|
|
|
33598
33634
|
def __init__(
|
|
33599
33635
|
self,
|
|
33600
33636
|
doc_id: str = None,
|
|
33637
|
+
model_name: str = None,
|
|
33601
33638
|
prompt: str = None,
|
|
33602
33639
|
session_id: str = None,
|
|
33603
33640
|
workspace_id: str = None,
|
|
33604
33641
|
):
|
|
33605
33642
|
# This parameter is required.
|
|
33606
33643
|
self.doc_id = doc_id
|
|
33644
|
+
self.model_name = model_name
|
|
33607
33645
|
self.prompt = prompt
|
|
33608
33646
|
# This parameter is required.
|
|
33609
33647
|
self.session_id = session_id
|
|
@@ -33621,6 +33659,8 @@ class RunDocSmartCardRequest(TeaModel):
|
|
|
33621
33659
|
result = dict()
|
|
33622
33660
|
if self.doc_id is not None:
|
|
33623
33661
|
result['DocId'] = self.doc_id
|
|
33662
|
+
if self.model_name is not None:
|
|
33663
|
+
result['ModelName'] = self.model_name
|
|
33624
33664
|
if self.prompt is not None:
|
|
33625
33665
|
result['Prompt'] = self.prompt
|
|
33626
33666
|
if self.session_id is not None:
|
|
@@ -33633,6 +33673,8 @@ class RunDocSmartCardRequest(TeaModel):
|
|
|
33633
33673
|
m = m or dict()
|
|
33634
33674
|
if m.get('DocId') is not None:
|
|
33635
33675
|
self.doc_id = m.get('DocId')
|
|
33676
|
+
if m.get('ModelName') is not None:
|
|
33677
|
+
self.model_name = m.get('ModelName')
|
|
33636
33678
|
if m.get('Prompt') is not None:
|
|
33637
33679
|
self.prompt = m.get('Prompt')
|
|
33638
33680
|
if m.get('SessionId') is not None:
|
|
@@ -34223,6 +34265,7 @@ class RunDocTranslationRequest(TeaModel):
|
|
|
34223
34265
|
self,
|
|
34224
34266
|
clean_cache: bool = None,
|
|
34225
34267
|
doc_id: str = None,
|
|
34268
|
+
model_name: str = None,
|
|
34226
34269
|
recommend_content: str = None,
|
|
34227
34270
|
session_id: str = None,
|
|
34228
34271
|
trans_type: str = None,
|
|
@@ -34230,6 +34273,7 @@ class RunDocTranslationRequest(TeaModel):
|
|
|
34230
34273
|
):
|
|
34231
34274
|
self.clean_cache = clean_cache
|
|
34232
34275
|
self.doc_id = doc_id
|
|
34276
|
+
self.model_name = model_name
|
|
34233
34277
|
self.recommend_content = recommend_content
|
|
34234
34278
|
# This parameter is required.
|
|
34235
34279
|
self.session_id = session_id
|
|
@@ -34250,6 +34294,8 @@ class RunDocTranslationRequest(TeaModel):
|
|
|
34250
34294
|
result['CleanCache'] = self.clean_cache
|
|
34251
34295
|
if self.doc_id is not None:
|
|
34252
34296
|
result['DocId'] = self.doc_id
|
|
34297
|
+
if self.model_name is not None:
|
|
34298
|
+
result['ModelName'] = self.model_name
|
|
34253
34299
|
if self.recommend_content is not None:
|
|
34254
34300
|
result['RecommendContent'] = self.recommend_content
|
|
34255
34301
|
if self.session_id is not None:
|
|
@@ -34266,6 +34312,8 @@ class RunDocTranslationRequest(TeaModel):
|
|
|
34266
34312
|
self.clean_cache = m.get('CleanCache')
|
|
34267
34313
|
if m.get('DocId') is not None:
|
|
34268
34314
|
self.doc_id = m.get('DocId')
|
|
34315
|
+
if m.get('ModelName') is not None:
|
|
34316
|
+
self.model_name = m.get('ModelName')
|
|
34269
34317
|
if m.get('RecommendContent') is not None:
|
|
34270
34318
|
self.recommend_content = m.get('RecommendContent')
|
|
34271
34319
|
if m.get('SessionId') is not None:
|
|
@@ -34532,6 +34580,7 @@ class RunDocTranslationResponse(TeaModel):
|
|
|
34532
34580
|
class RunDocWashingRequest(TeaModel):
|
|
34533
34581
|
def __init__(
|
|
34534
34582
|
self,
|
|
34583
|
+
model_id: str = None,
|
|
34535
34584
|
prompt: str = None,
|
|
34536
34585
|
reference_content: str = None,
|
|
34537
34586
|
session_id: str = None,
|
|
@@ -34541,6 +34590,7 @@ class RunDocWashingRequest(TeaModel):
|
|
|
34541
34590
|
writing_type_name: str = None,
|
|
34542
34591
|
writing_type_ref_doc: str = None,
|
|
34543
34592
|
):
|
|
34593
|
+
self.model_id = model_id
|
|
34544
34594
|
self.prompt = prompt
|
|
34545
34595
|
# This parameter is required.
|
|
34546
34596
|
self.reference_content = reference_content
|
|
@@ -34561,6 +34611,8 @@ class RunDocWashingRequest(TeaModel):
|
|
|
34561
34611
|
return _map
|
|
34562
34612
|
|
|
34563
34613
|
result = dict()
|
|
34614
|
+
if self.model_id is not None:
|
|
34615
|
+
result['ModelId'] = self.model_id
|
|
34564
34616
|
if self.prompt is not None:
|
|
34565
34617
|
result['Prompt'] = self.prompt
|
|
34566
34618
|
if self.reference_content is not None:
|
|
@@ -34581,6 +34633,8 @@ class RunDocWashingRequest(TeaModel):
|
|
|
34581
34633
|
|
|
34582
34634
|
def from_map(self, m: dict = None):
|
|
34583
34635
|
m = m or dict()
|
|
34636
|
+
if m.get('ModelId') is not None:
|
|
34637
|
+
self.model_id = m.get('ModelId')
|
|
34584
34638
|
if m.get('Prompt') is not None:
|
|
34585
34639
|
self.prompt = m.get('Prompt')
|
|
34586
34640
|
if m.get('ReferenceContent') is not None:
|
|
@@ -35150,11 +35204,13 @@ class RunGenerateQuestionsRequest(TeaModel):
|
|
|
35150
35204
|
def __init__(
|
|
35151
35205
|
self,
|
|
35152
35206
|
doc_id: str = None,
|
|
35207
|
+
model_name: str = None,
|
|
35153
35208
|
reference_content: str = None,
|
|
35154
35209
|
session_id: str = None,
|
|
35155
35210
|
workspace_id: str = None,
|
|
35156
35211
|
):
|
|
35157
35212
|
self.doc_id = doc_id
|
|
35213
|
+
self.model_name = model_name
|
|
35158
35214
|
self.reference_content = reference_content
|
|
35159
35215
|
self.session_id = session_id
|
|
35160
35216
|
# This parameter is required.
|
|
@@ -35171,6 +35227,8 @@ class RunGenerateQuestionsRequest(TeaModel):
|
|
|
35171
35227
|
result = dict()
|
|
35172
35228
|
if self.doc_id is not None:
|
|
35173
35229
|
result['DocId'] = self.doc_id
|
|
35230
|
+
if self.model_name is not None:
|
|
35231
|
+
result['ModelName'] = self.model_name
|
|
35174
35232
|
if self.reference_content is not None:
|
|
35175
35233
|
result['ReferenceContent'] = self.reference_content
|
|
35176
35234
|
if self.session_id is not None:
|
|
@@ -35183,6 +35241,8 @@ class RunGenerateQuestionsRequest(TeaModel):
|
|
|
35183
35241
|
m = m or dict()
|
|
35184
35242
|
if m.get('DocId') is not None:
|
|
35185
35243
|
self.doc_id = m.get('DocId')
|
|
35244
|
+
if m.get('ModelName') is not None:
|
|
35245
|
+
self.model_name = m.get('ModelName')
|
|
35186
35246
|
if m.get('ReferenceContent') is not None:
|
|
35187
35247
|
self.reference_content = m.get('ReferenceContent')
|
|
35188
35248
|
if m.get('SessionId') is not None:
|
|
@@ -35449,12 +35509,14 @@ class RunHotwordRequest(TeaModel):
|
|
|
35449
35509
|
def __init__(
|
|
35450
35510
|
self,
|
|
35451
35511
|
doc_id: str = None,
|
|
35512
|
+
model_name: str = None,
|
|
35452
35513
|
prompt: str = None,
|
|
35453
35514
|
reference_content: str = None,
|
|
35454
35515
|
session_id: str = None,
|
|
35455
35516
|
workspace_id: str = None,
|
|
35456
35517
|
):
|
|
35457
35518
|
self.doc_id = doc_id
|
|
35519
|
+
self.model_name = model_name
|
|
35458
35520
|
self.prompt = prompt
|
|
35459
35521
|
self.reference_content = reference_content
|
|
35460
35522
|
self.session_id = session_id
|
|
@@ -35472,6 +35534,8 @@ class RunHotwordRequest(TeaModel):
|
|
|
35472
35534
|
result = dict()
|
|
35473
35535
|
if self.doc_id is not None:
|
|
35474
35536
|
result['DocId'] = self.doc_id
|
|
35537
|
+
if self.model_name is not None:
|
|
35538
|
+
result['ModelName'] = self.model_name
|
|
35475
35539
|
if self.prompt is not None:
|
|
35476
35540
|
result['Prompt'] = self.prompt
|
|
35477
35541
|
if self.reference_content is not None:
|
|
@@ -35486,6 +35550,8 @@ class RunHotwordRequest(TeaModel):
|
|
|
35486
35550
|
m = m or dict()
|
|
35487
35551
|
if m.get('DocId') is not None:
|
|
35488
35552
|
self.doc_id = m.get('DocId')
|
|
35553
|
+
if m.get('ModelName') is not None:
|
|
35554
|
+
self.model_name = m.get('ModelName')
|
|
35489
35555
|
if m.get('Prompt') is not None:
|
|
35490
35556
|
self.prompt = m.get('Prompt')
|
|
35491
35557
|
if m.get('ReferenceContent') is not None:
|
|
@@ -36113,6 +36179,7 @@ class RunMultiDocIntroductionRequest(TeaModel):
|
|
|
36113
36179
|
self,
|
|
36114
36180
|
doc_ids: List[str] = None,
|
|
36115
36181
|
key_point_prompt: str = None,
|
|
36182
|
+
model_name: str = None,
|
|
36116
36183
|
session_id: str = None,
|
|
36117
36184
|
summary_prompt: str = None,
|
|
36118
36185
|
workspace_id: str = None,
|
|
@@ -36120,6 +36187,7 @@ class RunMultiDocIntroductionRequest(TeaModel):
|
|
|
36120
36187
|
# This parameter is required.
|
|
36121
36188
|
self.doc_ids = doc_ids
|
|
36122
36189
|
self.key_point_prompt = key_point_prompt
|
|
36190
|
+
self.model_name = model_name
|
|
36123
36191
|
# This parameter is required.
|
|
36124
36192
|
self.session_id = session_id
|
|
36125
36193
|
self.summary_prompt = summary_prompt
|
|
@@ -36139,6 +36207,8 @@ class RunMultiDocIntroductionRequest(TeaModel):
|
|
|
36139
36207
|
result['DocIds'] = self.doc_ids
|
|
36140
36208
|
if self.key_point_prompt is not None:
|
|
36141
36209
|
result['KeyPointPrompt'] = self.key_point_prompt
|
|
36210
|
+
if self.model_name is not None:
|
|
36211
|
+
result['ModelName'] = self.model_name
|
|
36142
36212
|
if self.session_id is not None:
|
|
36143
36213
|
result['SessionId'] = self.session_id
|
|
36144
36214
|
if self.summary_prompt is not None:
|
|
@@ -36153,6 +36223,8 @@ class RunMultiDocIntroductionRequest(TeaModel):
|
|
|
36153
36223
|
self.doc_ids = m.get('DocIds')
|
|
36154
36224
|
if m.get('KeyPointPrompt') is not None:
|
|
36155
36225
|
self.key_point_prompt = m.get('KeyPointPrompt')
|
|
36226
|
+
if m.get('ModelName') is not None:
|
|
36227
|
+
self.model_name = m.get('ModelName')
|
|
36156
36228
|
if m.get('SessionId') is not None:
|
|
36157
36229
|
self.session_id = m.get('SessionId')
|
|
36158
36230
|
if m.get('SummaryPrompt') is not None:
|
|
@@ -36167,6 +36239,7 @@ class RunMultiDocIntroductionShrinkRequest(TeaModel):
|
|
|
36167
36239
|
self,
|
|
36168
36240
|
doc_ids_shrink: str = None,
|
|
36169
36241
|
key_point_prompt: str = None,
|
|
36242
|
+
model_name: str = None,
|
|
36170
36243
|
session_id: str = None,
|
|
36171
36244
|
summary_prompt: str = None,
|
|
36172
36245
|
workspace_id: str = None,
|
|
@@ -36174,6 +36247,7 @@ class RunMultiDocIntroductionShrinkRequest(TeaModel):
|
|
|
36174
36247
|
# This parameter is required.
|
|
36175
36248
|
self.doc_ids_shrink = doc_ids_shrink
|
|
36176
36249
|
self.key_point_prompt = key_point_prompt
|
|
36250
|
+
self.model_name = model_name
|
|
36177
36251
|
# This parameter is required.
|
|
36178
36252
|
self.session_id = session_id
|
|
36179
36253
|
self.summary_prompt = summary_prompt
|
|
@@ -36193,6 +36267,8 @@ class RunMultiDocIntroductionShrinkRequest(TeaModel):
|
|
|
36193
36267
|
result['DocIds'] = self.doc_ids_shrink
|
|
36194
36268
|
if self.key_point_prompt is not None:
|
|
36195
36269
|
result['KeyPointPrompt'] = self.key_point_prompt
|
|
36270
|
+
if self.model_name is not None:
|
|
36271
|
+
result['ModelName'] = self.model_name
|
|
36196
36272
|
if self.session_id is not None:
|
|
36197
36273
|
result['SessionId'] = self.session_id
|
|
36198
36274
|
if self.summary_prompt is not None:
|
|
@@ -36207,6 +36283,8 @@ class RunMultiDocIntroductionShrinkRequest(TeaModel):
|
|
|
36207
36283
|
self.doc_ids_shrink = m.get('DocIds')
|
|
36208
36284
|
if m.get('KeyPointPrompt') is not None:
|
|
36209
36285
|
self.key_point_prompt = m.get('KeyPointPrompt')
|
|
36286
|
+
if m.get('ModelName') is not None:
|
|
36287
|
+
self.model_name = m.get('ModelName')
|
|
36210
36288
|
if m.get('SessionId') is not None:
|
|
36211
36289
|
self.session_id = m.get('SessionId')
|
|
36212
36290
|
if m.get('SummaryPrompt') is not None:
|
|
@@ -36911,12 +36989,14 @@ class RunSearchGenerationRequestChatConfig(TeaModel):
|
|
|
36911
36989
|
def __init__(
|
|
36912
36990
|
self,
|
|
36913
36991
|
enable_thinking: bool = None,
|
|
36992
|
+
exclude_generate_options: List[str] = None,
|
|
36914
36993
|
generate_level: str = None,
|
|
36915
36994
|
generate_technology: str = None,
|
|
36916
36995
|
search_models: List[str] = None,
|
|
36917
36996
|
search_param: RunSearchGenerationRequestChatConfigSearchParam = None,
|
|
36918
36997
|
):
|
|
36919
36998
|
self.enable_thinking = enable_thinking
|
|
36999
|
+
self.exclude_generate_options = exclude_generate_options
|
|
36920
37000
|
self.generate_level = generate_level
|
|
36921
37001
|
self.generate_technology = generate_technology
|
|
36922
37002
|
self.search_models = search_models
|
|
@@ -36934,6 +37014,8 @@ class RunSearchGenerationRequestChatConfig(TeaModel):
|
|
|
36934
37014
|
result = dict()
|
|
36935
37015
|
if self.enable_thinking is not None:
|
|
36936
37016
|
result['EnableThinking'] = self.enable_thinking
|
|
37017
|
+
if self.exclude_generate_options is not None:
|
|
37018
|
+
result['ExcludeGenerateOptions'] = self.exclude_generate_options
|
|
36937
37019
|
if self.generate_level is not None:
|
|
36938
37020
|
result['GenerateLevel'] = self.generate_level
|
|
36939
37021
|
if self.generate_technology is not None:
|
|
@@ -36948,6 +37030,8 @@ class RunSearchGenerationRequestChatConfig(TeaModel):
|
|
|
36948
37030
|
m = m or dict()
|
|
36949
37031
|
if m.get('EnableThinking') is not None:
|
|
36950
37032
|
self.enable_thinking = m.get('EnableThinking')
|
|
37033
|
+
if m.get('ExcludeGenerateOptions') is not None:
|
|
37034
|
+
self.exclude_generate_options = m.get('ExcludeGenerateOptions')
|
|
36951
37035
|
if m.get('GenerateLevel') is not None:
|
|
36952
37036
|
self.generate_level = m.get('GenerateLevel')
|
|
36953
37037
|
if m.get('GenerateTechnology') is not None:
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
__version__ = '1.30.0'
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|