alibabacloud-quanmiaolightapp20240801 2.6.0__tar.gz → 2.6.2__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (17) hide show
  1. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/ChangeLog.md +16 -0
  2. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/PKG-INFO +1 -1
  3. alibabacloud_quanmiaolightapp20240801-2.6.2/alibabacloud_quanmiaolightapp20240801/__init__.py +1 -0
  4. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801/client.py +32 -0
  5. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801/models.py +430 -1
  6. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801.egg-info/PKG-INFO +1 -1
  7. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/setup.py +1 -1
  8. alibabacloud_quanmiaolightapp20240801-2.6.0/alibabacloud_quanmiaolightapp20240801/__init__.py +0 -1
  9. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/LICENSE +0 -0
  10. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/MANIFEST.in +0 -0
  11. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/README-CN.md +0 -0
  12. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/README.md +0 -0
  13. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801.egg-info/SOURCES.txt +0 -0
  14. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801.egg-info/dependency_links.txt +0 -0
  15. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801.egg-info/requires.txt +0 -0
  16. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801.egg-info/top_level.txt +0 -0
  17. {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/setup.cfg +0 -0
@@ -1,3 +1,19 @@
1
+ 2025-03-13 Version: 2.6.1
2
+ - Update API GetVideoAnalysisTask: update response param.
3
+ - Update API RunVideoAnalysis: add param faceIdentitySimilarityMinScore.
4
+ - Update API RunVideoAnalysis: add param textProcessTasks.
5
+ - Update API RunVideoAnalysis: add param videoShotFaceIdentityCount.
6
+ - Update API RunVideoAnalysis: update response param.
7
+ - Update API SubmitVideoAnalysisTask: add param faceIdentitySimilarityMinScore.
8
+ - Update API SubmitVideoAnalysisTask: add param textProcessTasks.
9
+ - Update API SubmitVideoAnalysisTask: add param videoShotFaceIdentityCount.
10
+
11
+
12
+ 2025-03-10 Version: 2.6.0
13
+ - Support API GetTagMiningAnalysisTask.
14
+ - Support API SubmitTagMiningAnalysisTask.
15
+
16
+
1
17
  2025-01-23 Version: 2.5.1
2
18
  - Update API RunStyleWriting: add param processStage.
3
19
  - Update API RunStyleWriting: add param useSearch.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: alibabacloud_quanmiaolightapp20240801
3
- Version: 2.6.0
3
+ Version: 2.6.2
4
4
  Summary: Alibaba Cloud QuanMiaoLightApp (20240801) SDK Library for Python
5
5
  Home-page: https://github.com/aliyun/alibabacloud-python-sdk
6
6
  Author: Alibaba Cloud SDK
@@ -2385,9 +2385,13 @@ class Client(OpenApiClient):
2385
2385
  request.frame_sample_method_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.frame_sample_method, 'frameSampleMethod', 'json')
2386
2386
  if not UtilClient.is_unset(tmp_req.generate_options):
2387
2387
  request.generate_options_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.generate_options, 'generateOptions', 'json')
2388
+ if not UtilClient.is_unset(tmp_req.text_process_tasks):
2389
+ request.text_process_tasks_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.text_process_tasks, 'textProcessTasks', 'json')
2388
2390
  if not UtilClient.is_unset(tmp_req.video_roles):
2389
2391
  request.video_roles_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.video_roles, 'videoRoles', 'json')
2390
2392
  body = {}
2393
+ if not UtilClient.is_unset(request.face_identity_similarity_min_score):
2394
+ body['faceIdentitySimilarityMinScore'] = request.face_identity_similarity_min_score
2391
2395
  if not UtilClient.is_unset(request.frame_sample_method_shrink):
2392
2396
  body['frameSampleMethod'] = request.frame_sample_method_shrink
2393
2397
  if not UtilClient.is_unset(request.generate_options_shrink):
@@ -2406,6 +2410,8 @@ class Client(OpenApiClient):
2406
2410
  body['snapshotInterval'] = request.snapshot_interval
2407
2411
  if not UtilClient.is_unset(request.task_id):
2408
2412
  body['taskId'] = request.task_id
2413
+ if not UtilClient.is_unset(request.text_process_tasks_shrink):
2414
+ body['textProcessTasks'] = request.text_process_tasks_shrink
2409
2415
  if not UtilClient.is_unset(request.video_extra_info):
2410
2416
  body['videoExtraInfo'] = request.video_extra_info
2411
2417
  if not UtilClient.is_unset(request.video_model_custom_prompt_template):
@@ -2414,6 +2420,8 @@ class Client(OpenApiClient):
2414
2420
  body['videoModelId'] = request.video_model_id
2415
2421
  if not UtilClient.is_unset(request.video_roles_shrink):
2416
2422
  body['videoRoles'] = request.video_roles_shrink
2423
+ if not UtilClient.is_unset(request.video_shot_face_identity_count):
2424
+ body['videoShotFaceIdentityCount'] = request.video_shot_face_identity_count
2417
2425
  if not UtilClient.is_unset(request.video_url):
2418
2426
  body['videoUrl'] = request.video_url
2419
2427
  req = open_api_models.OpenApiRequest(
@@ -2464,9 +2472,13 @@ class Client(OpenApiClient):
2464
2472
  request.frame_sample_method_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.frame_sample_method, 'frameSampleMethod', 'json')
2465
2473
  if not UtilClient.is_unset(tmp_req.generate_options):
2466
2474
  request.generate_options_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.generate_options, 'generateOptions', 'json')
2475
+ if not UtilClient.is_unset(tmp_req.text_process_tasks):
2476
+ request.text_process_tasks_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.text_process_tasks, 'textProcessTasks', 'json')
2467
2477
  if not UtilClient.is_unset(tmp_req.video_roles):
2468
2478
  request.video_roles_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.video_roles, 'videoRoles', 'json')
2469
2479
  body = {}
2480
+ if not UtilClient.is_unset(request.face_identity_similarity_min_score):
2481
+ body['faceIdentitySimilarityMinScore'] = request.face_identity_similarity_min_score
2470
2482
  if not UtilClient.is_unset(request.frame_sample_method_shrink):
2471
2483
  body['frameSampleMethod'] = request.frame_sample_method_shrink
2472
2484
  if not UtilClient.is_unset(request.generate_options_shrink):
@@ -2485,6 +2497,8 @@ class Client(OpenApiClient):
2485
2497
  body['snapshotInterval'] = request.snapshot_interval
2486
2498
  if not UtilClient.is_unset(request.task_id):
2487
2499
  body['taskId'] = request.task_id
2500
+ if not UtilClient.is_unset(request.text_process_tasks_shrink):
2501
+ body['textProcessTasks'] = request.text_process_tasks_shrink
2488
2502
  if not UtilClient.is_unset(request.video_extra_info):
2489
2503
  body['videoExtraInfo'] = request.video_extra_info
2490
2504
  if not UtilClient.is_unset(request.video_model_custom_prompt_template):
@@ -2493,6 +2507,8 @@ class Client(OpenApiClient):
2493
2507
  body['videoModelId'] = request.video_model_id
2494
2508
  if not UtilClient.is_unset(request.video_roles_shrink):
2495
2509
  body['videoRoles'] = request.video_roles_shrink
2510
+ if not UtilClient.is_unset(request.video_shot_face_identity_count):
2511
+ body['videoShotFaceIdentityCount'] = request.video_shot_face_identity_count
2496
2512
  if not UtilClient.is_unset(request.video_url):
2497
2513
  body['videoUrl'] = request.video_url
2498
2514
  req = open_api_models.OpenApiRequest(
@@ -2733,9 +2749,13 @@ class Client(OpenApiClient):
2733
2749
  request.frame_sample_method_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.frame_sample_method, 'frameSampleMethod', 'json')
2734
2750
  if not UtilClient.is_unset(tmp_req.generate_options):
2735
2751
  request.generate_options_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.generate_options, 'generateOptions', 'json')
2752
+ if not UtilClient.is_unset(tmp_req.text_process_tasks):
2753
+ request.text_process_tasks_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.text_process_tasks, 'textProcessTasks', 'json')
2736
2754
  if not UtilClient.is_unset(tmp_req.video_roles):
2737
2755
  request.video_roles_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.video_roles, 'videoRoles', 'json')
2738
2756
  body = {}
2757
+ if not UtilClient.is_unset(request.face_identity_similarity_min_score):
2758
+ body['faceIdentitySimilarityMinScore'] = request.face_identity_similarity_min_score
2739
2759
  if not UtilClient.is_unset(request.frame_sample_method_shrink):
2740
2760
  body['frameSampleMethod'] = request.frame_sample_method_shrink
2741
2761
  if not UtilClient.is_unset(request.generate_options_shrink):
@@ -2750,6 +2770,8 @@ class Client(OpenApiClient):
2750
2770
  body['modelId'] = request.model_id
2751
2771
  if not UtilClient.is_unset(request.snapshot_interval):
2752
2772
  body['snapshotInterval'] = request.snapshot_interval
2773
+ if not UtilClient.is_unset(request.text_process_tasks_shrink):
2774
+ body['textProcessTasks'] = request.text_process_tasks_shrink
2753
2775
  if not UtilClient.is_unset(request.video_extra_info):
2754
2776
  body['videoExtraInfo'] = request.video_extra_info
2755
2777
  if not UtilClient.is_unset(request.video_model_custom_prompt_template):
@@ -2758,6 +2780,8 @@ class Client(OpenApiClient):
2758
2780
  body['videoModelId'] = request.video_model_id
2759
2781
  if not UtilClient.is_unset(request.video_roles_shrink):
2760
2782
  body['videoRoles'] = request.video_roles_shrink
2783
+ if not UtilClient.is_unset(request.video_shot_face_identity_count):
2784
+ body['videoShotFaceIdentityCount'] = request.video_shot_face_identity_count
2761
2785
  if not UtilClient.is_unset(request.video_url):
2762
2786
  body['videoUrl'] = request.video_url
2763
2787
  req = open_api_models.OpenApiRequest(
@@ -2808,9 +2832,13 @@ class Client(OpenApiClient):
2808
2832
  request.frame_sample_method_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.frame_sample_method, 'frameSampleMethod', 'json')
2809
2833
  if not UtilClient.is_unset(tmp_req.generate_options):
2810
2834
  request.generate_options_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.generate_options, 'generateOptions', 'json')
2835
+ if not UtilClient.is_unset(tmp_req.text_process_tasks):
2836
+ request.text_process_tasks_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.text_process_tasks, 'textProcessTasks', 'json')
2811
2837
  if not UtilClient.is_unset(tmp_req.video_roles):
2812
2838
  request.video_roles_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.video_roles, 'videoRoles', 'json')
2813
2839
  body = {}
2840
+ if not UtilClient.is_unset(request.face_identity_similarity_min_score):
2841
+ body['faceIdentitySimilarityMinScore'] = request.face_identity_similarity_min_score
2814
2842
  if not UtilClient.is_unset(request.frame_sample_method_shrink):
2815
2843
  body['frameSampleMethod'] = request.frame_sample_method_shrink
2816
2844
  if not UtilClient.is_unset(request.generate_options_shrink):
@@ -2825,6 +2853,8 @@ class Client(OpenApiClient):
2825
2853
  body['modelId'] = request.model_id
2826
2854
  if not UtilClient.is_unset(request.snapshot_interval):
2827
2855
  body['snapshotInterval'] = request.snapshot_interval
2856
+ if not UtilClient.is_unset(request.text_process_tasks_shrink):
2857
+ body['textProcessTasks'] = request.text_process_tasks_shrink
2828
2858
  if not UtilClient.is_unset(request.video_extra_info):
2829
2859
  body['videoExtraInfo'] = request.video_extra_info
2830
2860
  if not UtilClient.is_unset(request.video_model_custom_prompt_template):
@@ -2833,6 +2863,8 @@ class Client(OpenApiClient):
2833
2863
  body['videoModelId'] = request.video_model_id
2834
2864
  if not UtilClient.is_unset(request.video_roles_shrink):
2835
2865
  body['videoRoles'] = request.video_roles_shrink
2866
+ if not UtilClient.is_unset(request.video_shot_face_identity_count):
2867
+ body['videoShotFaceIdentityCount'] = request.video_shot_face_identity_count
2836
2868
  if not UtilClient.is_unset(request.video_url):
2837
2869
  body['videoUrl'] = request.video_url
2838
2870
  req = open_api_models.OpenApiRequest(
@@ -1466,10 +1466,18 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult(TeaMo
1466
1466
  def __init__(
1467
1467
  self,
1468
1468
  generate_finished: bool = None,
1469
+ index: int = None,
1470
+ model_id: str = None,
1471
+ model_reduce: bool = None,
1472
+ reason_text: str = None,
1469
1473
  text: str = None,
1470
1474
  usage: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResultUsage = None,
1471
1475
  ):
1472
1476
  self.generate_finished = generate_finished
1477
+ self.index = index
1478
+ self.model_id = model_id
1479
+ self.model_reduce = model_reduce
1480
+ self.reason_text = reason_text
1473
1481
  self.text = text
1474
1482
  self.usage = usage
1475
1483
 
@@ -1485,6 +1493,14 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult(TeaMo
1485
1493
  result = dict()
1486
1494
  if self.generate_finished is not None:
1487
1495
  result['generateFinished'] = self.generate_finished
1496
+ if self.index is not None:
1497
+ result['index'] = self.index
1498
+ if self.model_id is not None:
1499
+ result['modelId'] = self.model_id
1500
+ if self.model_reduce is not None:
1501
+ result['modelReduce'] = self.model_reduce
1502
+ if self.reason_text is not None:
1503
+ result['reasonText'] = self.reason_text
1488
1504
  if self.text is not None:
1489
1505
  result['text'] = self.text
1490
1506
  if self.usage is not None:
@@ -1495,6 +1511,14 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult(TeaMo
1495
1511
  m = m or dict()
1496
1512
  if m.get('generateFinished') is not None:
1497
1513
  self.generate_finished = m.get('generateFinished')
1514
+ if m.get('index') is not None:
1515
+ self.index = m.get('index')
1516
+ if m.get('modelId') is not None:
1517
+ self.model_id = m.get('modelId')
1518
+ if m.get('modelReduce') is not None:
1519
+ self.model_reduce = m.get('modelReduce')
1520
+ if m.get('reasonText') is not None:
1521
+ self.reason_text = m.get('reasonText')
1498
1522
  if m.get('text') is not None:
1499
1523
  self.text = m.get('text')
1500
1524
  if m.get('usage') is not None:
@@ -1503,6 +1527,104 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult(TeaMo
1503
1527
  return self
1504
1528
 
1505
1529
 
1530
+ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResultsUsage(TeaModel):
1531
+ def __init__(
1532
+ self,
1533
+ input_tokens: int = None,
1534
+ output_tokens: int = None,
1535
+ total_tokens: int = None,
1536
+ ):
1537
+ self.input_tokens = input_tokens
1538
+ self.output_tokens = output_tokens
1539
+ self.total_tokens = total_tokens
1540
+
1541
+ def validate(self):
1542
+ pass
1543
+
1544
+ def to_map(self):
1545
+ _map = super().to_map()
1546
+ if _map is not None:
1547
+ return _map
1548
+
1549
+ result = dict()
1550
+ if self.input_tokens is not None:
1551
+ result['inputTokens'] = self.input_tokens
1552
+ if self.output_tokens is not None:
1553
+ result['outputTokens'] = self.output_tokens
1554
+ if self.total_tokens is not None:
1555
+ result['totalTokens'] = self.total_tokens
1556
+ return result
1557
+
1558
+ def from_map(self, m: dict = None):
1559
+ m = m or dict()
1560
+ if m.get('inputTokens') is not None:
1561
+ self.input_tokens = m.get('inputTokens')
1562
+ if m.get('outputTokens') is not None:
1563
+ self.output_tokens = m.get('outputTokens')
1564
+ if m.get('totalTokens') is not None:
1565
+ self.total_tokens = m.get('totalTokens')
1566
+ return self
1567
+
1568
+
1569
+ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResults(TeaModel):
1570
+ def __init__(
1571
+ self,
1572
+ generate_finished: bool = None,
1573
+ index: int = None,
1574
+ model_id: str = None,
1575
+ reason_text: str = None,
1576
+ text: str = None,
1577
+ usage: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResultsUsage = None,
1578
+ ):
1579
+ self.generate_finished = generate_finished
1580
+ self.index = index
1581
+ self.model_id = model_id
1582
+ self.reason_text = reason_text
1583
+ self.text = text
1584
+ self.usage = usage
1585
+
1586
+ def validate(self):
1587
+ if self.usage:
1588
+ self.usage.validate()
1589
+
1590
+ def to_map(self):
1591
+ _map = super().to_map()
1592
+ if _map is not None:
1593
+ return _map
1594
+
1595
+ result = dict()
1596
+ if self.generate_finished is not None:
1597
+ result['generateFinished'] = self.generate_finished
1598
+ if self.index is not None:
1599
+ result['index'] = self.index
1600
+ if self.model_id is not None:
1601
+ result['modelId'] = self.model_id
1602
+ if self.reason_text is not None:
1603
+ result['reasonText'] = self.reason_text
1604
+ if self.text is not None:
1605
+ result['text'] = self.text
1606
+ if self.usage is not None:
1607
+ result['usage'] = self.usage.to_map()
1608
+ return result
1609
+
1610
+ def from_map(self, m: dict = None):
1611
+ m = m or dict()
1612
+ if m.get('generateFinished') is not None:
1613
+ self.generate_finished = m.get('generateFinished')
1614
+ if m.get('index') is not None:
1615
+ self.index = m.get('index')
1616
+ if m.get('modelId') is not None:
1617
+ self.model_id = m.get('modelId')
1618
+ if m.get('reasonText') is not None:
1619
+ self.reason_text = m.get('reasonText')
1620
+ if m.get('text') is not None:
1621
+ self.text = m.get('text')
1622
+ if m.get('usage') is not None:
1623
+ temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResultsUsage()
1624
+ self.usage = temp_model.from_map(m['usage'])
1625
+ return self
1626
+
1627
+
1506
1628
  class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoMindMappingGenerateResultUsage(TeaModel):
1507
1629
  def __init__(
1508
1630
  self,
@@ -1794,6 +1916,7 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
1794
1916
  video_analysis_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoAnalysisResult = None,
1795
1917
  video_caption_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoCaptionResult = None,
1796
1918
  video_generate_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult = None,
1919
+ video_generate_results: List[GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResults] = None,
1797
1920
  video_mind_mapping_generate_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoMindMappingGenerateResult = None,
1798
1921
  video_title_generate_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoTitleGenerateResult = None,
1799
1922
  ):
@@ -1801,6 +1924,7 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
1801
1924
  self.video_analysis_result = video_analysis_result
1802
1925
  self.video_caption_result = video_caption_result
1803
1926
  self.video_generate_result = video_generate_result
1927
+ self.video_generate_results = video_generate_results
1804
1928
  self.video_mind_mapping_generate_result = video_mind_mapping_generate_result
1805
1929
  self.video_title_generate_result = video_title_generate_result
1806
1930
 
@@ -1811,6 +1935,10 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
1811
1935
  self.video_caption_result.validate()
1812
1936
  if self.video_generate_result:
1813
1937
  self.video_generate_result.validate()
1938
+ if self.video_generate_results:
1939
+ for k in self.video_generate_results:
1940
+ if k:
1941
+ k.validate()
1814
1942
  if self.video_mind_mapping_generate_result:
1815
1943
  self.video_mind_mapping_generate_result.validate()
1816
1944
  if self.video_title_generate_result:
@@ -1830,6 +1958,10 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
1830
1958
  result['videoCaptionResult'] = self.video_caption_result.to_map()
1831
1959
  if self.video_generate_result is not None:
1832
1960
  result['videoGenerateResult'] = self.video_generate_result.to_map()
1961
+ result['videoGenerateResults'] = []
1962
+ if self.video_generate_results is not None:
1963
+ for k in self.video_generate_results:
1964
+ result['videoGenerateResults'].append(k.to_map() if k else None)
1833
1965
  if self.video_mind_mapping_generate_result is not None:
1834
1966
  result['videoMindMappingGenerateResult'] = self.video_mind_mapping_generate_result.to_map()
1835
1967
  if self.video_title_generate_result is not None:
@@ -1849,6 +1981,11 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
1849
1981
  if m.get('videoGenerateResult') is not None:
1850
1982
  temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult()
1851
1983
  self.video_generate_result = temp_model.from_map(m['videoGenerateResult'])
1984
+ self.video_generate_results = []
1985
+ if m.get('videoGenerateResults') is not None:
1986
+ for k in m.get('videoGenerateResults'):
1987
+ temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResults()
1988
+ self.video_generate_results.append(temp_model.from_map(k))
1852
1989
  if m.get('videoMindMappingGenerateResult') is not None:
1853
1990
  temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoMindMappingGenerateResult()
1854
1991
  self.video_mind_mapping_generate_result = temp_model.from_map(m['videoMindMappingGenerateResult'])
@@ -3430,7 +3567,6 @@ class RunHotTopicSummaryRequestStepForCustomSummaryStyleConfig(TeaModel):
3430
3567
  summary_prompt: str = None,
3431
3568
  ):
3432
3569
  self.summary_image_count = summary_image_count
3433
- # This parameter is required.
3434
3570
  self.summary_model = summary_model
3435
3571
  self.summary_prompt = summary_prompt
3436
3572
 
@@ -6916,6 +7052,45 @@ class RunVideoAnalysisRequestFrameSampleMethod(TeaModel):
6916
7052
  return self
6917
7053
 
6918
7054
 
7055
+ class RunVideoAnalysisRequestTextProcessTasks(TeaModel):
7056
+ def __init__(
7057
+ self,
7058
+ model_custom_prompt_template: str = None,
7059
+ model_custom_prompt_template_id: str = None,
7060
+ model_id: str = None,
7061
+ ):
7062
+ self.model_custom_prompt_template = model_custom_prompt_template
7063
+ self.model_custom_prompt_template_id = model_custom_prompt_template_id
7064
+ self.model_id = model_id
7065
+
7066
+ def validate(self):
7067
+ pass
7068
+
7069
+ def to_map(self):
7070
+ _map = super().to_map()
7071
+ if _map is not None:
7072
+ return _map
7073
+
7074
+ result = dict()
7075
+ if self.model_custom_prompt_template is not None:
7076
+ result['modelCustomPromptTemplate'] = self.model_custom_prompt_template
7077
+ if self.model_custom_prompt_template_id is not None:
7078
+ result['modelCustomPromptTemplateId'] = self.model_custom_prompt_template_id
7079
+ if self.model_id is not None:
7080
+ result['modelId'] = self.model_id
7081
+ return result
7082
+
7083
+ def from_map(self, m: dict = None):
7084
+ m = m or dict()
7085
+ if m.get('modelCustomPromptTemplate') is not None:
7086
+ self.model_custom_prompt_template = m.get('modelCustomPromptTemplate')
7087
+ if m.get('modelCustomPromptTemplateId') is not None:
7088
+ self.model_custom_prompt_template_id = m.get('modelCustomPromptTemplateId')
7089
+ if m.get('modelId') is not None:
7090
+ self.model_id = m.get('modelId')
7091
+ return self
7092
+
7093
+
6919
7094
  class RunVideoAnalysisRequestVideoRoles(TeaModel):
6920
7095
  def __init__(
6921
7096
  self,
@@ -6958,6 +7133,7 @@ class RunVideoAnalysisRequestVideoRoles(TeaModel):
6958
7133
  class RunVideoAnalysisRequest(TeaModel):
6959
7134
  def __init__(
6960
7135
  self,
7136
+ face_identity_similarity_min_score: float = None,
6961
7137
  frame_sample_method: RunVideoAnalysisRequestFrameSampleMethod = None,
6962
7138
  generate_options: List[str] = None,
6963
7139
  language: str = None,
@@ -6967,12 +7143,15 @@ class RunVideoAnalysisRequest(TeaModel):
6967
7143
  original_session_id: str = None,
6968
7144
  snapshot_interval: float = None,
6969
7145
  task_id: str = None,
7146
+ text_process_tasks: List[RunVideoAnalysisRequestTextProcessTasks] = None,
6970
7147
  video_extra_info: str = None,
6971
7148
  video_model_custom_prompt_template: str = None,
6972
7149
  video_model_id: str = None,
6973
7150
  video_roles: List[RunVideoAnalysisRequestVideoRoles] = None,
7151
+ video_shot_face_identity_count: int = None,
6974
7152
  video_url: str = None,
6975
7153
  ):
7154
+ self.face_identity_similarity_min_score = face_identity_similarity_min_score
6976
7155
  self.frame_sample_method = frame_sample_method
6977
7156
  self.generate_options = generate_options
6978
7157
  self.language = language
@@ -6982,15 +7161,21 @@ class RunVideoAnalysisRequest(TeaModel):
6982
7161
  self.original_session_id = original_session_id
6983
7162
  self.snapshot_interval = snapshot_interval
6984
7163
  self.task_id = task_id
7164
+ self.text_process_tasks = text_process_tasks
6985
7165
  self.video_extra_info = video_extra_info
6986
7166
  self.video_model_custom_prompt_template = video_model_custom_prompt_template
6987
7167
  self.video_model_id = video_model_id
6988
7168
  self.video_roles = video_roles
7169
+ self.video_shot_face_identity_count = video_shot_face_identity_count
6989
7170
  self.video_url = video_url
6990
7171
 
6991
7172
  def validate(self):
6992
7173
  if self.frame_sample_method:
6993
7174
  self.frame_sample_method.validate()
7175
+ if self.text_process_tasks:
7176
+ for k in self.text_process_tasks:
7177
+ if k:
7178
+ k.validate()
6994
7179
  if self.video_roles:
6995
7180
  for k in self.video_roles:
6996
7181
  if k:
@@ -7002,6 +7187,8 @@ class RunVideoAnalysisRequest(TeaModel):
7002
7187
  return _map
7003
7188
 
7004
7189
  result = dict()
7190
+ if self.face_identity_similarity_min_score is not None:
7191
+ result['faceIdentitySimilarityMinScore'] = self.face_identity_similarity_min_score
7005
7192
  if self.frame_sample_method is not None:
7006
7193
  result['frameSampleMethod'] = self.frame_sample_method.to_map()
7007
7194
  if self.generate_options is not None:
@@ -7020,6 +7207,10 @@ class RunVideoAnalysisRequest(TeaModel):
7020
7207
  result['snapshotInterval'] = self.snapshot_interval
7021
7208
  if self.task_id is not None:
7022
7209
  result['taskId'] = self.task_id
7210
+ result['textProcessTasks'] = []
7211
+ if self.text_process_tasks is not None:
7212
+ for k in self.text_process_tasks:
7213
+ result['textProcessTasks'].append(k.to_map() if k else None)
7023
7214
  if self.video_extra_info is not None:
7024
7215
  result['videoExtraInfo'] = self.video_extra_info
7025
7216
  if self.video_model_custom_prompt_template is not None:
@@ -7030,12 +7221,16 @@ class RunVideoAnalysisRequest(TeaModel):
7030
7221
  if self.video_roles is not None:
7031
7222
  for k in self.video_roles:
7032
7223
  result['videoRoles'].append(k.to_map() if k else None)
7224
+ if self.video_shot_face_identity_count is not None:
7225
+ result['videoShotFaceIdentityCount'] = self.video_shot_face_identity_count
7033
7226
  if self.video_url is not None:
7034
7227
  result['videoUrl'] = self.video_url
7035
7228
  return result
7036
7229
 
7037
7230
  def from_map(self, m: dict = None):
7038
7231
  m = m or dict()
7232
+ if m.get('faceIdentitySimilarityMinScore') is not None:
7233
+ self.face_identity_similarity_min_score = m.get('faceIdentitySimilarityMinScore')
7039
7234
  if m.get('frameSampleMethod') is not None:
7040
7235
  temp_model = RunVideoAnalysisRequestFrameSampleMethod()
7041
7236
  self.frame_sample_method = temp_model.from_map(m['frameSampleMethod'])
@@ -7055,6 +7250,11 @@ class RunVideoAnalysisRequest(TeaModel):
7055
7250
  self.snapshot_interval = m.get('snapshotInterval')
7056
7251
  if m.get('taskId') is not None:
7057
7252
  self.task_id = m.get('taskId')
7253
+ self.text_process_tasks = []
7254
+ if m.get('textProcessTasks') is not None:
7255
+ for k in m.get('textProcessTasks'):
7256
+ temp_model = RunVideoAnalysisRequestTextProcessTasks()
7257
+ self.text_process_tasks.append(temp_model.from_map(k))
7058
7258
  if m.get('videoExtraInfo') is not None:
7059
7259
  self.video_extra_info = m.get('videoExtraInfo')
7060
7260
  if m.get('videoModelCustomPromptTemplate') is not None:
@@ -7066,6 +7266,8 @@ class RunVideoAnalysisRequest(TeaModel):
7066
7266
  for k in m.get('videoRoles'):
7067
7267
  temp_model = RunVideoAnalysisRequestVideoRoles()
7068
7268
  self.video_roles.append(temp_model.from_map(k))
7269
+ if m.get('videoShotFaceIdentityCount') is not None:
7270
+ self.video_shot_face_identity_count = m.get('videoShotFaceIdentityCount')
7069
7271
  if m.get('videoUrl') is not None:
7070
7272
  self.video_url = m.get('videoUrl')
7071
7273
  return self
@@ -7074,6 +7276,7 @@ class RunVideoAnalysisRequest(TeaModel):
7074
7276
  class RunVideoAnalysisShrinkRequest(TeaModel):
7075
7277
  def __init__(
7076
7278
  self,
7279
+ face_identity_similarity_min_score: float = None,
7077
7280
  frame_sample_method_shrink: str = None,
7078
7281
  generate_options_shrink: str = None,
7079
7282
  language: str = None,
@@ -7083,12 +7286,15 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
7083
7286
  original_session_id: str = None,
7084
7287
  snapshot_interval: float = None,
7085
7288
  task_id: str = None,
7289
+ text_process_tasks_shrink: str = None,
7086
7290
  video_extra_info: str = None,
7087
7291
  video_model_custom_prompt_template: str = None,
7088
7292
  video_model_id: str = None,
7089
7293
  video_roles_shrink: str = None,
7294
+ video_shot_face_identity_count: int = None,
7090
7295
  video_url: str = None,
7091
7296
  ):
7297
+ self.face_identity_similarity_min_score = face_identity_similarity_min_score
7092
7298
  self.frame_sample_method_shrink = frame_sample_method_shrink
7093
7299
  self.generate_options_shrink = generate_options_shrink
7094
7300
  self.language = language
@@ -7098,10 +7304,12 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
7098
7304
  self.original_session_id = original_session_id
7099
7305
  self.snapshot_interval = snapshot_interval
7100
7306
  self.task_id = task_id
7307
+ self.text_process_tasks_shrink = text_process_tasks_shrink
7101
7308
  self.video_extra_info = video_extra_info
7102
7309
  self.video_model_custom_prompt_template = video_model_custom_prompt_template
7103
7310
  self.video_model_id = video_model_id
7104
7311
  self.video_roles_shrink = video_roles_shrink
7312
+ self.video_shot_face_identity_count = video_shot_face_identity_count
7105
7313
  self.video_url = video_url
7106
7314
 
7107
7315
  def validate(self):
@@ -7113,6 +7321,8 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
7113
7321
  return _map
7114
7322
 
7115
7323
  result = dict()
7324
+ if self.face_identity_similarity_min_score is not None:
7325
+ result['faceIdentitySimilarityMinScore'] = self.face_identity_similarity_min_score
7116
7326
  if self.frame_sample_method_shrink is not None:
7117
7327
  result['frameSampleMethod'] = self.frame_sample_method_shrink
7118
7328
  if self.generate_options_shrink is not None:
@@ -7131,6 +7341,8 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
7131
7341
  result['snapshotInterval'] = self.snapshot_interval
7132
7342
  if self.task_id is not None:
7133
7343
  result['taskId'] = self.task_id
7344
+ if self.text_process_tasks_shrink is not None:
7345
+ result['textProcessTasks'] = self.text_process_tasks_shrink
7134
7346
  if self.video_extra_info is not None:
7135
7347
  result['videoExtraInfo'] = self.video_extra_info
7136
7348
  if self.video_model_custom_prompt_template is not None:
@@ -7139,12 +7351,16 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
7139
7351
  result['videoModelId'] = self.video_model_id
7140
7352
  if self.video_roles_shrink is not None:
7141
7353
  result['videoRoles'] = self.video_roles_shrink
7354
+ if self.video_shot_face_identity_count is not None:
7355
+ result['videoShotFaceIdentityCount'] = self.video_shot_face_identity_count
7142
7356
  if self.video_url is not None:
7143
7357
  result['videoUrl'] = self.video_url
7144
7358
  return result
7145
7359
 
7146
7360
  def from_map(self, m: dict = None):
7147
7361
  m = m or dict()
7362
+ if m.get('faceIdentitySimilarityMinScore') is not None:
7363
+ self.face_identity_similarity_min_score = m.get('faceIdentitySimilarityMinScore')
7148
7364
  if m.get('frameSampleMethod') is not None:
7149
7365
  self.frame_sample_method_shrink = m.get('frameSampleMethod')
7150
7366
  if m.get('generateOptions') is not None:
@@ -7163,6 +7379,8 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
7163
7379
  self.snapshot_interval = m.get('snapshotInterval')
7164
7380
  if m.get('taskId') is not None:
7165
7381
  self.task_id = m.get('taskId')
7382
+ if m.get('textProcessTasks') is not None:
7383
+ self.text_process_tasks_shrink = m.get('textProcessTasks')
7166
7384
  if m.get('videoExtraInfo') is not None:
7167
7385
  self.video_extra_info = m.get('videoExtraInfo')
7168
7386
  if m.get('videoModelCustomPromptTemplate') is not None:
@@ -7171,6 +7389,8 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
7171
7389
  self.video_model_id = m.get('videoModelId')
7172
7390
  if m.get('videoRoles') is not None:
7173
7391
  self.video_roles_shrink = m.get('videoRoles')
7392
+ if m.get('videoShotFaceIdentityCount') is not None:
7393
+ self.video_shot_face_identity_count = m.get('videoShotFaceIdentityCount')
7174
7394
  if m.get('videoUrl') is not None:
7175
7395
  self.video_url = m.get('videoUrl')
7176
7396
  return self
@@ -7514,14 +7734,18 @@ class RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult(TeaModel):
7514
7734
  def __init__(
7515
7735
  self,
7516
7736
  generate_finished: bool = None,
7737
+ index: int = None,
7517
7738
  model_id: str = None,
7518
7739
  model_reduce: bool = None,
7740
+ reason_text: str = None,
7519
7741
  text: str = None,
7520
7742
  usage: RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResultUsage = None,
7521
7743
  ):
7522
7744
  self.generate_finished = generate_finished
7745
+ self.index = index
7523
7746
  self.model_id = model_id
7524
7747
  self.model_reduce = model_reduce
7748
+ self.reason_text = reason_text
7525
7749
  self.text = text
7526
7750
  self.usage = usage
7527
7751
 
@@ -7537,10 +7761,14 @@ class RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult(TeaModel):
7537
7761
  result = dict()
7538
7762
  if self.generate_finished is not None:
7539
7763
  result['generateFinished'] = self.generate_finished
7764
+ if self.index is not None:
7765
+ result['index'] = self.index
7540
7766
  if self.model_id is not None:
7541
7767
  result['modelId'] = self.model_id
7542
7768
  if self.model_reduce is not None:
7543
7769
  result['modelReduce'] = self.model_reduce
7770
+ if self.reason_text is not None:
7771
+ result['reasonText'] = self.reason_text
7544
7772
  if self.text is not None:
7545
7773
  result['text'] = self.text
7546
7774
  if self.usage is not None:
@@ -7551,10 +7779,14 @@ class RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult(TeaModel):
7551
7779
  m = m or dict()
7552
7780
  if m.get('generateFinished') is not None:
7553
7781
  self.generate_finished = m.get('generateFinished')
7782
+ if m.get('index') is not None:
7783
+ self.index = m.get('index')
7554
7784
  if m.get('modelId') is not None:
7555
7785
  self.model_id = m.get('modelId')
7556
7786
  if m.get('modelReduce') is not None:
7557
7787
  self.model_reduce = m.get('modelReduce')
7788
+ if m.get('reasonText') is not None:
7789
+ self.reason_text = m.get('reasonText')
7558
7790
  if m.get('text') is not None:
7559
7791
  self.text = m.get('text')
7560
7792
  if m.get('usage') is not None:
@@ -7563,6 +7795,104 @@ class RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult(TeaModel):
7563
7795
  return self
7564
7796
 
7565
7797
 
7798
+ class RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResultsUsage(TeaModel):
7799
+ def __init__(
7800
+ self,
7801
+ input_tokens: int = None,
7802
+ output_tokens: int = None,
7803
+ total_tokens: int = None,
7804
+ ):
7805
+ self.input_tokens = input_tokens
7806
+ self.output_tokens = output_tokens
7807
+ self.total_tokens = total_tokens
7808
+
7809
+ def validate(self):
7810
+ pass
7811
+
7812
+ def to_map(self):
7813
+ _map = super().to_map()
7814
+ if _map is not None:
7815
+ return _map
7816
+
7817
+ result = dict()
7818
+ if self.input_tokens is not None:
7819
+ result['inputTokens'] = self.input_tokens
7820
+ if self.output_tokens is not None:
7821
+ result['outputTokens'] = self.output_tokens
7822
+ if self.total_tokens is not None:
7823
+ result['totalTokens'] = self.total_tokens
7824
+ return result
7825
+
7826
+ def from_map(self, m: dict = None):
7827
+ m = m or dict()
7828
+ if m.get('inputTokens') is not None:
7829
+ self.input_tokens = m.get('inputTokens')
7830
+ if m.get('outputTokens') is not None:
7831
+ self.output_tokens = m.get('outputTokens')
7832
+ if m.get('totalTokens') is not None:
7833
+ self.total_tokens = m.get('totalTokens')
7834
+ return self
7835
+
7836
+
7837
+ class RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResults(TeaModel):
7838
+ def __init__(
7839
+ self,
7840
+ generate_finished: bool = None,
7841
+ index: int = None,
7842
+ model_id: str = None,
7843
+ reason_text: str = None,
7844
+ text: str = None,
7845
+ usage: RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResultsUsage = None,
7846
+ ):
7847
+ self.generate_finished = generate_finished
7848
+ self.index = index
7849
+ self.model_id = model_id
7850
+ self.reason_text = reason_text
7851
+ self.text = text
7852
+ self.usage = usage
7853
+
7854
+ def validate(self):
7855
+ if self.usage:
7856
+ self.usage.validate()
7857
+
7858
+ def to_map(self):
7859
+ _map = super().to_map()
7860
+ if _map is not None:
7861
+ return _map
7862
+
7863
+ result = dict()
7864
+ if self.generate_finished is not None:
7865
+ result['generateFinished'] = self.generate_finished
7866
+ if self.index is not None:
7867
+ result['index'] = self.index
7868
+ if self.model_id is not None:
7869
+ result['modelId'] = self.model_id
7870
+ if self.reason_text is not None:
7871
+ result['reasonText'] = self.reason_text
7872
+ if self.text is not None:
7873
+ result['text'] = self.text
7874
+ if self.usage is not None:
7875
+ result['usage'] = self.usage.to_map()
7876
+ return result
7877
+
7878
+ def from_map(self, m: dict = None):
7879
+ m = m or dict()
7880
+ if m.get('generateFinished') is not None:
7881
+ self.generate_finished = m.get('generateFinished')
7882
+ if m.get('index') is not None:
7883
+ self.index = m.get('index')
7884
+ if m.get('modelId') is not None:
7885
+ self.model_id = m.get('modelId')
7886
+ if m.get('reasonText') is not None:
7887
+ self.reason_text = m.get('reasonText')
7888
+ if m.get('text') is not None:
7889
+ self.text = m.get('text')
7890
+ if m.get('usage') is not None:
7891
+ temp_model = RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResultsUsage()
7892
+ self.usage = temp_model.from_map(m['usage'])
7893
+ return self
7894
+
7895
+
7566
7896
  class RunVideoAnalysisResponseBodyPayloadOutputVideoMindMappingGenerateResultUsage(TeaModel):
7567
7897
  def __init__(
7568
7898
  self,
@@ -7999,6 +8329,7 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
7999
8329
  video_analysis_result: RunVideoAnalysisResponseBodyPayloadOutputVideoAnalysisResult = None,
8000
8330
  video_caption_result: RunVideoAnalysisResponseBodyPayloadOutputVideoCaptionResult = None,
8001
8331
  video_generate_result: RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult = None,
8332
+ video_generate_results: List[RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResults] = None,
8002
8333
  video_mind_mapping_generate_result: RunVideoAnalysisResponseBodyPayloadOutputVideoMindMappingGenerateResult = None,
8003
8334
  video_shot_snapshot_result: RunVideoAnalysisResponseBodyPayloadOutputVideoShotSnapshotResult = None,
8004
8335
  video_title_generate_result: RunVideoAnalysisResponseBodyPayloadOutputVideoTitleGenerateResult = None,
@@ -8007,6 +8338,7 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
8007
8338
  self.video_analysis_result = video_analysis_result
8008
8339
  self.video_caption_result = video_caption_result
8009
8340
  self.video_generate_result = video_generate_result
8341
+ self.video_generate_results = video_generate_results
8010
8342
  self.video_mind_mapping_generate_result = video_mind_mapping_generate_result
8011
8343
  self.video_shot_snapshot_result = video_shot_snapshot_result
8012
8344
  self.video_title_generate_result = video_title_generate_result
@@ -8018,6 +8350,10 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
8018
8350
  self.video_caption_result.validate()
8019
8351
  if self.video_generate_result:
8020
8352
  self.video_generate_result.validate()
8353
+ if self.video_generate_results:
8354
+ for k in self.video_generate_results:
8355
+ if k:
8356
+ k.validate()
8021
8357
  if self.video_mind_mapping_generate_result:
8022
8358
  self.video_mind_mapping_generate_result.validate()
8023
8359
  if self.video_shot_snapshot_result:
@@ -8039,6 +8375,10 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
8039
8375
  result['videoCaptionResult'] = self.video_caption_result.to_map()
8040
8376
  if self.video_generate_result is not None:
8041
8377
  result['videoGenerateResult'] = self.video_generate_result.to_map()
8378
+ result['videoGenerateResults'] = []
8379
+ if self.video_generate_results is not None:
8380
+ for k in self.video_generate_results:
8381
+ result['videoGenerateResults'].append(k.to_map() if k else None)
8042
8382
  if self.video_mind_mapping_generate_result is not None:
8043
8383
  result['videoMindMappingGenerateResult'] = self.video_mind_mapping_generate_result.to_map()
8044
8384
  if self.video_shot_snapshot_result is not None:
@@ -8060,6 +8400,11 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
8060
8400
  if m.get('videoGenerateResult') is not None:
8061
8401
  temp_model = RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult()
8062
8402
  self.video_generate_result = temp_model.from_map(m['videoGenerateResult'])
8403
+ self.video_generate_results = []
8404
+ if m.get('videoGenerateResults') is not None:
8405
+ for k in m.get('videoGenerateResults'):
8406
+ temp_model = RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResults()
8407
+ self.video_generate_results.append(temp_model.from_map(k))
8063
8408
  if m.get('videoMindMappingGenerateResult') is not None:
8064
8409
  temp_model = RunVideoAnalysisResponseBodyPayloadOutputVideoMindMappingGenerateResult()
8065
8410
  self.video_mind_mapping_generate_result = temp_model.from_map(m['videoMindMappingGenerateResult'])
@@ -8580,6 +8925,45 @@ class SubmitVideoAnalysisTaskRequestFrameSampleMethod(TeaModel):
8580
8925
  return self
8581
8926
 
8582
8927
 
8928
+ class SubmitVideoAnalysisTaskRequestTextProcessTasks(TeaModel):
8929
+ def __init__(
8930
+ self,
8931
+ model_custom_prompt_template: str = None,
8932
+ model_custom_prompt_template_id: str = None,
8933
+ model_id: str = None,
8934
+ ):
8935
+ self.model_custom_prompt_template = model_custom_prompt_template
8936
+ self.model_custom_prompt_template_id = model_custom_prompt_template_id
8937
+ self.model_id = model_id
8938
+
8939
+ def validate(self):
8940
+ pass
8941
+
8942
+ def to_map(self):
8943
+ _map = super().to_map()
8944
+ if _map is not None:
8945
+ return _map
8946
+
8947
+ result = dict()
8948
+ if self.model_custom_prompt_template is not None:
8949
+ result['modelCustomPromptTemplate'] = self.model_custom_prompt_template
8950
+ if self.model_custom_prompt_template_id is not None:
8951
+ result['modelCustomPromptTemplateId'] = self.model_custom_prompt_template_id
8952
+ if self.model_id is not None:
8953
+ result['modelId'] = self.model_id
8954
+ return result
8955
+
8956
+ def from_map(self, m: dict = None):
8957
+ m = m or dict()
8958
+ if m.get('modelCustomPromptTemplate') is not None:
8959
+ self.model_custom_prompt_template = m.get('modelCustomPromptTemplate')
8960
+ if m.get('modelCustomPromptTemplateId') is not None:
8961
+ self.model_custom_prompt_template_id = m.get('modelCustomPromptTemplateId')
8962
+ if m.get('modelId') is not None:
8963
+ self.model_id = m.get('modelId')
8964
+ return self
8965
+
8966
+
8583
8967
  class SubmitVideoAnalysisTaskRequestVideoRoles(TeaModel):
8584
8968
  def __init__(
8585
8969
  self,
@@ -8622,6 +9006,7 @@ class SubmitVideoAnalysisTaskRequestVideoRoles(TeaModel):
8622
9006
  class SubmitVideoAnalysisTaskRequest(TeaModel):
8623
9007
  def __init__(
8624
9008
  self,
9009
+ face_identity_similarity_min_score: float = None,
8625
9010
  frame_sample_method: SubmitVideoAnalysisTaskRequestFrameSampleMethod = None,
8626
9011
  generate_options: List[str] = None,
8627
9012
  language: str = None,
@@ -8629,12 +9014,15 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
8629
9014
  model_custom_prompt_template_id: str = None,
8630
9015
  model_id: str = None,
8631
9016
  snapshot_interval: float = None,
9017
+ text_process_tasks: List[SubmitVideoAnalysisTaskRequestTextProcessTasks] = None,
8632
9018
  video_extra_info: str = None,
8633
9019
  video_model_custom_prompt_template: str = None,
8634
9020
  video_model_id: str = None,
8635
9021
  video_roles: List[SubmitVideoAnalysisTaskRequestVideoRoles] = None,
9022
+ video_shot_face_identity_count: int = None,
8636
9023
  video_url: str = None,
8637
9024
  ):
9025
+ self.face_identity_similarity_min_score = face_identity_similarity_min_score
8638
9026
  self.frame_sample_method = frame_sample_method
8639
9027
  self.generate_options = generate_options
8640
9028
  self.language = language
@@ -8642,16 +9030,22 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
8642
9030
  self.model_custom_prompt_template_id = model_custom_prompt_template_id
8643
9031
  self.model_id = model_id
8644
9032
  self.snapshot_interval = snapshot_interval
9033
+ self.text_process_tasks = text_process_tasks
8645
9034
  self.video_extra_info = video_extra_info
8646
9035
  self.video_model_custom_prompt_template = video_model_custom_prompt_template
8647
9036
  self.video_model_id = video_model_id
8648
9037
  self.video_roles = video_roles
9038
+ self.video_shot_face_identity_count = video_shot_face_identity_count
8649
9039
  # This parameter is required.
8650
9040
  self.video_url = video_url
8651
9041
 
8652
9042
  def validate(self):
8653
9043
  if self.frame_sample_method:
8654
9044
  self.frame_sample_method.validate()
9045
+ if self.text_process_tasks:
9046
+ for k in self.text_process_tasks:
9047
+ if k:
9048
+ k.validate()
8655
9049
  if self.video_roles:
8656
9050
  for k in self.video_roles:
8657
9051
  if k:
@@ -8663,6 +9057,8 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
8663
9057
  return _map
8664
9058
 
8665
9059
  result = dict()
9060
+ if self.face_identity_similarity_min_score is not None:
9061
+ result['faceIdentitySimilarityMinScore'] = self.face_identity_similarity_min_score
8666
9062
  if self.frame_sample_method is not None:
8667
9063
  result['frameSampleMethod'] = self.frame_sample_method.to_map()
8668
9064
  if self.generate_options is not None:
@@ -8677,6 +9073,10 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
8677
9073
  result['modelId'] = self.model_id
8678
9074
  if self.snapshot_interval is not None:
8679
9075
  result['snapshotInterval'] = self.snapshot_interval
9076
+ result['textProcessTasks'] = []
9077
+ if self.text_process_tasks is not None:
9078
+ for k in self.text_process_tasks:
9079
+ result['textProcessTasks'].append(k.to_map() if k else None)
8680
9080
  if self.video_extra_info is not None:
8681
9081
  result['videoExtraInfo'] = self.video_extra_info
8682
9082
  if self.video_model_custom_prompt_template is not None:
@@ -8687,12 +9087,16 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
8687
9087
  if self.video_roles is not None:
8688
9088
  for k in self.video_roles:
8689
9089
  result['videoRoles'].append(k.to_map() if k else None)
9090
+ if self.video_shot_face_identity_count is not None:
9091
+ result['videoShotFaceIdentityCount'] = self.video_shot_face_identity_count
8690
9092
  if self.video_url is not None:
8691
9093
  result['videoUrl'] = self.video_url
8692
9094
  return result
8693
9095
 
8694
9096
  def from_map(self, m: dict = None):
8695
9097
  m = m or dict()
9098
+ if m.get('faceIdentitySimilarityMinScore') is not None:
9099
+ self.face_identity_similarity_min_score = m.get('faceIdentitySimilarityMinScore')
8696
9100
  if m.get('frameSampleMethod') is not None:
8697
9101
  temp_model = SubmitVideoAnalysisTaskRequestFrameSampleMethod()
8698
9102
  self.frame_sample_method = temp_model.from_map(m['frameSampleMethod'])
@@ -8708,6 +9112,11 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
8708
9112
  self.model_id = m.get('modelId')
8709
9113
  if m.get('snapshotInterval') is not None:
8710
9114
  self.snapshot_interval = m.get('snapshotInterval')
9115
+ self.text_process_tasks = []
9116
+ if m.get('textProcessTasks') is not None:
9117
+ for k in m.get('textProcessTasks'):
9118
+ temp_model = SubmitVideoAnalysisTaskRequestTextProcessTasks()
9119
+ self.text_process_tasks.append(temp_model.from_map(k))
8711
9120
  if m.get('videoExtraInfo') is not None:
8712
9121
  self.video_extra_info = m.get('videoExtraInfo')
8713
9122
  if m.get('videoModelCustomPromptTemplate') is not None:
@@ -8719,6 +9128,8 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
8719
9128
  for k in m.get('videoRoles'):
8720
9129
  temp_model = SubmitVideoAnalysisTaskRequestVideoRoles()
8721
9130
  self.video_roles.append(temp_model.from_map(k))
9131
+ if m.get('videoShotFaceIdentityCount') is not None:
9132
+ self.video_shot_face_identity_count = m.get('videoShotFaceIdentityCount')
8722
9133
  if m.get('videoUrl') is not None:
8723
9134
  self.video_url = m.get('videoUrl')
8724
9135
  return self
@@ -8727,6 +9138,7 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
8727
9138
  class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
8728
9139
  def __init__(
8729
9140
  self,
9141
+ face_identity_similarity_min_score: float = None,
8730
9142
  frame_sample_method_shrink: str = None,
8731
9143
  generate_options_shrink: str = None,
8732
9144
  language: str = None,
@@ -8734,12 +9146,15 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
8734
9146
  model_custom_prompt_template_id: str = None,
8735
9147
  model_id: str = None,
8736
9148
  snapshot_interval: float = None,
9149
+ text_process_tasks_shrink: str = None,
8737
9150
  video_extra_info: str = None,
8738
9151
  video_model_custom_prompt_template: str = None,
8739
9152
  video_model_id: str = None,
8740
9153
  video_roles_shrink: str = None,
9154
+ video_shot_face_identity_count: int = None,
8741
9155
  video_url: str = None,
8742
9156
  ):
9157
+ self.face_identity_similarity_min_score = face_identity_similarity_min_score
8743
9158
  self.frame_sample_method_shrink = frame_sample_method_shrink
8744
9159
  self.generate_options_shrink = generate_options_shrink
8745
9160
  self.language = language
@@ -8747,10 +9162,12 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
8747
9162
  self.model_custom_prompt_template_id = model_custom_prompt_template_id
8748
9163
  self.model_id = model_id
8749
9164
  self.snapshot_interval = snapshot_interval
9165
+ self.text_process_tasks_shrink = text_process_tasks_shrink
8750
9166
  self.video_extra_info = video_extra_info
8751
9167
  self.video_model_custom_prompt_template = video_model_custom_prompt_template
8752
9168
  self.video_model_id = video_model_id
8753
9169
  self.video_roles_shrink = video_roles_shrink
9170
+ self.video_shot_face_identity_count = video_shot_face_identity_count
8754
9171
  # This parameter is required.
8755
9172
  self.video_url = video_url
8756
9173
 
@@ -8763,6 +9180,8 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
8763
9180
  return _map
8764
9181
 
8765
9182
  result = dict()
9183
+ if self.face_identity_similarity_min_score is not None:
9184
+ result['faceIdentitySimilarityMinScore'] = self.face_identity_similarity_min_score
8766
9185
  if self.frame_sample_method_shrink is not None:
8767
9186
  result['frameSampleMethod'] = self.frame_sample_method_shrink
8768
9187
  if self.generate_options_shrink is not None:
@@ -8777,6 +9196,8 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
8777
9196
  result['modelId'] = self.model_id
8778
9197
  if self.snapshot_interval is not None:
8779
9198
  result['snapshotInterval'] = self.snapshot_interval
9199
+ if self.text_process_tasks_shrink is not None:
9200
+ result['textProcessTasks'] = self.text_process_tasks_shrink
8780
9201
  if self.video_extra_info is not None:
8781
9202
  result['videoExtraInfo'] = self.video_extra_info
8782
9203
  if self.video_model_custom_prompt_template is not None:
@@ -8785,12 +9206,16 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
8785
9206
  result['videoModelId'] = self.video_model_id
8786
9207
  if self.video_roles_shrink is not None:
8787
9208
  result['videoRoles'] = self.video_roles_shrink
9209
+ if self.video_shot_face_identity_count is not None:
9210
+ result['videoShotFaceIdentityCount'] = self.video_shot_face_identity_count
8788
9211
  if self.video_url is not None:
8789
9212
  result['videoUrl'] = self.video_url
8790
9213
  return result
8791
9214
 
8792
9215
  def from_map(self, m: dict = None):
8793
9216
  m = m or dict()
9217
+ if m.get('faceIdentitySimilarityMinScore') is not None:
9218
+ self.face_identity_similarity_min_score = m.get('faceIdentitySimilarityMinScore')
8794
9219
  if m.get('frameSampleMethod') is not None:
8795
9220
  self.frame_sample_method_shrink = m.get('frameSampleMethod')
8796
9221
  if m.get('generateOptions') is not None:
@@ -8805,6 +9230,8 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
8805
9230
  self.model_id = m.get('modelId')
8806
9231
  if m.get('snapshotInterval') is not None:
8807
9232
  self.snapshot_interval = m.get('snapshotInterval')
9233
+ if m.get('textProcessTasks') is not None:
9234
+ self.text_process_tasks_shrink = m.get('textProcessTasks')
8808
9235
  if m.get('videoExtraInfo') is not None:
8809
9236
  self.video_extra_info = m.get('videoExtraInfo')
8810
9237
  if m.get('videoModelCustomPromptTemplate') is not None:
@@ -8813,6 +9240,8 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
8813
9240
  self.video_model_id = m.get('videoModelId')
8814
9241
  if m.get('videoRoles') is not None:
8815
9242
  self.video_roles_shrink = m.get('videoRoles')
9243
+ if m.get('videoShotFaceIdentityCount') is not None:
9244
+ self.video_shot_face_identity_count = m.get('videoShotFaceIdentityCount')
8816
9245
  if m.get('videoUrl') is not None:
8817
9246
  self.video_url = m.get('videoUrl')
8818
9247
  return self
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: alibabacloud-quanmiaolightapp20240801
3
- Version: 2.6.0
3
+ Version: 2.6.2
4
4
  Summary: Alibaba Cloud QuanMiaoLightApp (20240801) SDK Library for Python
5
5
  Home-page: https://github.com/aliyun/alibabacloud-python-sdk
6
6
  Author: Alibaba Cloud SDK
@@ -24,7 +24,7 @@ from setuptools import setup, find_packages
24
24
  """
25
25
  setup module for alibabacloud_quanmiaolightapp20240801.
26
26
 
27
- Created on 10/03/2025
27
+ Created on 17/03/2025
28
28
 
29
29
  @author: Alibaba Cloud SDK
30
30
  """