alibabacloud-quanmiaolightapp20240801 2.6.0__tar.gz → 2.6.2__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/ChangeLog.md +16 -0
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/PKG-INFO +1 -1
- alibabacloud_quanmiaolightapp20240801-2.6.2/alibabacloud_quanmiaolightapp20240801/__init__.py +1 -0
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801/client.py +32 -0
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801/models.py +430 -1
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801.egg-info/PKG-INFO +1 -1
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/setup.py +1 -1
- alibabacloud_quanmiaolightapp20240801-2.6.0/alibabacloud_quanmiaolightapp20240801/__init__.py +0 -1
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/LICENSE +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/MANIFEST.in +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/README-CN.md +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/README.md +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801.egg-info/SOURCES.txt +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801.egg-info/dependency_links.txt +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801.egg-info/requires.txt +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/alibabacloud_quanmiaolightapp20240801.egg-info/top_level.txt +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/setup.cfg +0 -0
|
@@ -1,3 +1,19 @@
|
|
|
1
|
+
2025-03-13 Version: 2.6.1
|
|
2
|
+
- Update API GetVideoAnalysisTask: update response param.
|
|
3
|
+
- Update API RunVideoAnalysis: add param faceIdentitySimilarityMinScore.
|
|
4
|
+
- Update API RunVideoAnalysis: add param textProcessTasks.
|
|
5
|
+
- Update API RunVideoAnalysis: add param videoShotFaceIdentityCount.
|
|
6
|
+
- Update API RunVideoAnalysis: update response param.
|
|
7
|
+
- Update API SubmitVideoAnalysisTask: add param faceIdentitySimilarityMinScore.
|
|
8
|
+
- Update API SubmitVideoAnalysisTask: add param textProcessTasks.
|
|
9
|
+
- Update API SubmitVideoAnalysisTask: add param videoShotFaceIdentityCount.
|
|
10
|
+
|
|
11
|
+
|
|
12
|
+
2025-03-10 Version: 2.6.0
|
|
13
|
+
- Support API GetTagMiningAnalysisTask.
|
|
14
|
+
- Support API SubmitTagMiningAnalysisTask.
|
|
15
|
+
|
|
16
|
+
|
|
1
17
|
2025-01-23 Version: 2.5.1
|
|
2
18
|
- Update API RunStyleWriting: add param processStage.
|
|
3
19
|
- Update API RunStyleWriting: add param useSearch.
|
{alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/PKG-INFO
RENAMED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.1
|
|
2
2
|
Name: alibabacloud_quanmiaolightapp20240801
|
|
3
|
-
Version: 2.6.
|
|
3
|
+
Version: 2.6.2
|
|
4
4
|
Summary: Alibaba Cloud QuanMiaoLightApp (20240801) SDK Library for Python
|
|
5
5
|
Home-page: https://github.com/aliyun/alibabacloud-python-sdk
|
|
6
6
|
Author: Alibaba Cloud SDK
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
__version__ = '2.6.2'
|
|
@@ -2385,9 +2385,13 @@ class Client(OpenApiClient):
|
|
|
2385
2385
|
request.frame_sample_method_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.frame_sample_method, 'frameSampleMethod', 'json')
|
|
2386
2386
|
if not UtilClient.is_unset(tmp_req.generate_options):
|
|
2387
2387
|
request.generate_options_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.generate_options, 'generateOptions', 'json')
|
|
2388
|
+
if not UtilClient.is_unset(tmp_req.text_process_tasks):
|
|
2389
|
+
request.text_process_tasks_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.text_process_tasks, 'textProcessTasks', 'json')
|
|
2388
2390
|
if not UtilClient.is_unset(tmp_req.video_roles):
|
|
2389
2391
|
request.video_roles_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.video_roles, 'videoRoles', 'json')
|
|
2390
2392
|
body = {}
|
|
2393
|
+
if not UtilClient.is_unset(request.face_identity_similarity_min_score):
|
|
2394
|
+
body['faceIdentitySimilarityMinScore'] = request.face_identity_similarity_min_score
|
|
2391
2395
|
if not UtilClient.is_unset(request.frame_sample_method_shrink):
|
|
2392
2396
|
body['frameSampleMethod'] = request.frame_sample_method_shrink
|
|
2393
2397
|
if not UtilClient.is_unset(request.generate_options_shrink):
|
|
@@ -2406,6 +2410,8 @@ class Client(OpenApiClient):
|
|
|
2406
2410
|
body['snapshotInterval'] = request.snapshot_interval
|
|
2407
2411
|
if not UtilClient.is_unset(request.task_id):
|
|
2408
2412
|
body['taskId'] = request.task_id
|
|
2413
|
+
if not UtilClient.is_unset(request.text_process_tasks_shrink):
|
|
2414
|
+
body['textProcessTasks'] = request.text_process_tasks_shrink
|
|
2409
2415
|
if not UtilClient.is_unset(request.video_extra_info):
|
|
2410
2416
|
body['videoExtraInfo'] = request.video_extra_info
|
|
2411
2417
|
if not UtilClient.is_unset(request.video_model_custom_prompt_template):
|
|
@@ -2414,6 +2420,8 @@ class Client(OpenApiClient):
|
|
|
2414
2420
|
body['videoModelId'] = request.video_model_id
|
|
2415
2421
|
if not UtilClient.is_unset(request.video_roles_shrink):
|
|
2416
2422
|
body['videoRoles'] = request.video_roles_shrink
|
|
2423
|
+
if not UtilClient.is_unset(request.video_shot_face_identity_count):
|
|
2424
|
+
body['videoShotFaceIdentityCount'] = request.video_shot_face_identity_count
|
|
2417
2425
|
if not UtilClient.is_unset(request.video_url):
|
|
2418
2426
|
body['videoUrl'] = request.video_url
|
|
2419
2427
|
req = open_api_models.OpenApiRequest(
|
|
@@ -2464,9 +2472,13 @@ class Client(OpenApiClient):
|
|
|
2464
2472
|
request.frame_sample_method_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.frame_sample_method, 'frameSampleMethod', 'json')
|
|
2465
2473
|
if not UtilClient.is_unset(tmp_req.generate_options):
|
|
2466
2474
|
request.generate_options_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.generate_options, 'generateOptions', 'json')
|
|
2475
|
+
if not UtilClient.is_unset(tmp_req.text_process_tasks):
|
|
2476
|
+
request.text_process_tasks_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.text_process_tasks, 'textProcessTasks', 'json')
|
|
2467
2477
|
if not UtilClient.is_unset(tmp_req.video_roles):
|
|
2468
2478
|
request.video_roles_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.video_roles, 'videoRoles', 'json')
|
|
2469
2479
|
body = {}
|
|
2480
|
+
if not UtilClient.is_unset(request.face_identity_similarity_min_score):
|
|
2481
|
+
body['faceIdentitySimilarityMinScore'] = request.face_identity_similarity_min_score
|
|
2470
2482
|
if not UtilClient.is_unset(request.frame_sample_method_shrink):
|
|
2471
2483
|
body['frameSampleMethod'] = request.frame_sample_method_shrink
|
|
2472
2484
|
if not UtilClient.is_unset(request.generate_options_shrink):
|
|
@@ -2485,6 +2497,8 @@ class Client(OpenApiClient):
|
|
|
2485
2497
|
body['snapshotInterval'] = request.snapshot_interval
|
|
2486
2498
|
if not UtilClient.is_unset(request.task_id):
|
|
2487
2499
|
body['taskId'] = request.task_id
|
|
2500
|
+
if not UtilClient.is_unset(request.text_process_tasks_shrink):
|
|
2501
|
+
body['textProcessTasks'] = request.text_process_tasks_shrink
|
|
2488
2502
|
if not UtilClient.is_unset(request.video_extra_info):
|
|
2489
2503
|
body['videoExtraInfo'] = request.video_extra_info
|
|
2490
2504
|
if not UtilClient.is_unset(request.video_model_custom_prompt_template):
|
|
@@ -2493,6 +2507,8 @@ class Client(OpenApiClient):
|
|
|
2493
2507
|
body['videoModelId'] = request.video_model_id
|
|
2494
2508
|
if not UtilClient.is_unset(request.video_roles_shrink):
|
|
2495
2509
|
body['videoRoles'] = request.video_roles_shrink
|
|
2510
|
+
if not UtilClient.is_unset(request.video_shot_face_identity_count):
|
|
2511
|
+
body['videoShotFaceIdentityCount'] = request.video_shot_face_identity_count
|
|
2496
2512
|
if not UtilClient.is_unset(request.video_url):
|
|
2497
2513
|
body['videoUrl'] = request.video_url
|
|
2498
2514
|
req = open_api_models.OpenApiRequest(
|
|
@@ -2733,9 +2749,13 @@ class Client(OpenApiClient):
|
|
|
2733
2749
|
request.frame_sample_method_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.frame_sample_method, 'frameSampleMethod', 'json')
|
|
2734
2750
|
if not UtilClient.is_unset(tmp_req.generate_options):
|
|
2735
2751
|
request.generate_options_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.generate_options, 'generateOptions', 'json')
|
|
2752
|
+
if not UtilClient.is_unset(tmp_req.text_process_tasks):
|
|
2753
|
+
request.text_process_tasks_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.text_process_tasks, 'textProcessTasks', 'json')
|
|
2736
2754
|
if not UtilClient.is_unset(tmp_req.video_roles):
|
|
2737
2755
|
request.video_roles_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.video_roles, 'videoRoles', 'json')
|
|
2738
2756
|
body = {}
|
|
2757
|
+
if not UtilClient.is_unset(request.face_identity_similarity_min_score):
|
|
2758
|
+
body['faceIdentitySimilarityMinScore'] = request.face_identity_similarity_min_score
|
|
2739
2759
|
if not UtilClient.is_unset(request.frame_sample_method_shrink):
|
|
2740
2760
|
body['frameSampleMethod'] = request.frame_sample_method_shrink
|
|
2741
2761
|
if not UtilClient.is_unset(request.generate_options_shrink):
|
|
@@ -2750,6 +2770,8 @@ class Client(OpenApiClient):
|
|
|
2750
2770
|
body['modelId'] = request.model_id
|
|
2751
2771
|
if not UtilClient.is_unset(request.snapshot_interval):
|
|
2752
2772
|
body['snapshotInterval'] = request.snapshot_interval
|
|
2773
|
+
if not UtilClient.is_unset(request.text_process_tasks_shrink):
|
|
2774
|
+
body['textProcessTasks'] = request.text_process_tasks_shrink
|
|
2753
2775
|
if not UtilClient.is_unset(request.video_extra_info):
|
|
2754
2776
|
body['videoExtraInfo'] = request.video_extra_info
|
|
2755
2777
|
if not UtilClient.is_unset(request.video_model_custom_prompt_template):
|
|
@@ -2758,6 +2780,8 @@ class Client(OpenApiClient):
|
|
|
2758
2780
|
body['videoModelId'] = request.video_model_id
|
|
2759
2781
|
if not UtilClient.is_unset(request.video_roles_shrink):
|
|
2760
2782
|
body['videoRoles'] = request.video_roles_shrink
|
|
2783
|
+
if not UtilClient.is_unset(request.video_shot_face_identity_count):
|
|
2784
|
+
body['videoShotFaceIdentityCount'] = request.video_shot_face_identity_count
|
|
2761
2785
|
if not UtilClient.is_unset(request.video_url):
|
|
2762
2786
|
body['videoUrl'] = request.video_url
|
|
2763
2787
|
req = open_api_models.OpenApiRequest(
|
|
@@ -2808,9 +2832,13 @@ class Client(OpenApiClient):
|
|
|
2808
2832
|
request.frame_sample_method_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.frame_sample_method, 'frameSampleMethod', 'json')
|
|
2809
2833
|
if not UtilClient.is_unset(tmp_req.generate_options):
|
|
2810
2834
|
request.generate_options_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.generate_options, 'generateOptions', 'json')
|
|
2835
|
+
if not UtilClient.is_unset(tmp_req.text_process_tasks):
|
|
2836
|
+
request.text_process_tasks_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.text_process_tasks, 'textProcessTasks', 'json')
|
|
2811
2837
|
if not UtilClient.is_unset(tmp_req.video_roles):
|
|
2812
2838
|
request.video_roles_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.video_roles, 'videoRoles', 'json')
|
|
2813
2839
|
body = {}
|
|
2840
|
+
if not UtilClient.is_unset(request.face_identity_similarity_min_score):
|
|
2841
|
+
body['faceIdentitySimilarityMinScore'] = request.face_identity_similarity_min_score
|
|
2814
2842
|
if not UtilClient.is_unset(request.frame_sample_method_shrink):
|
|
2815
2843
|
body['frameSampleMethod'] = request.frame_sample_method_shrink
|
|
2816
2844
|
if not UtilClient.is_unset(request.generate_options_shrink):
|
|
@@ -2825,6 +2853,8 @@ class Client(OpenApiClient):
|
|
|
2825
2853
|
body['modelId'] = request.model_id
|
|
2826
2854
|
if not UtilClient.is_unset(request.snapshot_interval):
|
|
2827
2855
|
body['snapshotInterval'] = request.snapshot_interval
|
|
2856
|
+
if not UtilClient.is_unset(request.text_process_tasks_shrink):
|
|
2857
|
+
body['textProcessTasks'] = request.text_process_tasks_shrink
|
|
2828
2858
|
if not UtilClient.is_unset(request.video_extra_info):
|
|
2829
2859
|
body['videoExtraInfo'] = request.video_extra_info
|
|
2830
2860
|
if not UtilClient.is_unset(request.video_model_custom_prompt_template):
|
|
@@ -2833,6 +2863,8 @@ class Client(OpenApiClient):
|
|
|
2833
2863
|
body['videoModelId'] = request.video_model_id
|
|
2834
2864
|
if not UtilClient.is_unset(request.video_roles_shrink):
|
|
2835
2865
|
body['videoRoles'] = request.video_roles_shrink
|
|
2866
|
+
if not UtilClient.is_unset(request.video_shot_face_identity_count):
|
|
2867
|
+
body['videoShotFaceIdentityCount'] = request.video_shot_face_identity_count
|
|
2836
2868
|
if not UtilClient.is_unset(request.video_url):
|
|
2837
2869
|
body['videoUrl'] = request.video_url
|
|
2838
2870
|
req = open_api_models.OpenApiRequest(
|
|
@@ -1466,10 +1466,18 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult(TeaMo
|
|
|
1466
1466
|
def __init__(
|
|
1467
1467
|
self,
|
|
1468
1468
|
generate_finished: bool = None,
|
|
1469
|
+
index: int = None,
|
|
1470
|
+
model_id: str = None,
|
|
1471
|
+
model_reduce: bool = None,
|
|
1472
|
+
reason_text: str = None,
|
|
1469
1473
|
text: str = None,
|
|
1470
1474
|
usage: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResultUsage = None,
|
|
1471
1475
|
):
|
|
1472
1476
|
self.generate_finished = generate_finished
|
|
1477
|
+
self.index = index
|
|
1478
|
+
self.model_id = model_id
|
|
1479
|
+
self.model_reduce = model_reduce
|
|
1480
|
+
self.reason_text = reason_text
|
|
1473
1481
|
self.text = text
|
|
1474
1482
|
self.usage = usage
|
|
1475
1483
|
|
|
@@ -1485,6 +1493,14 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult(TeaMo
|
|
|
1485
1493
|
result = dict()
|
|
1486
1494
|
if self.generate_finished is not None:
|
|
1487
1495
|
result['generateFinished'] = self.generate_finished
|
|
1496
|
+
if self.index is not None:
|
|
1497
|
+
result['index'] = self.index
|
|
1498
|
+
if self.model_id is not None:
|
|
1499
|
+
result['modelId'] = self.model_id
|
|
1500
|
+
if self.model_reduce is not None:
|
|
1501
|
+
result['modelReduce'] = self.model_reduce
|
|
1502
|
+
if self.reason_text is not None:
|
|
1503
|
+
result['reasonText'] = self.reason_text
|
|
1488
1504
|
if self.text is not None:
|
|
1489
1505
|
result['text'] = self.text
|
|
1490
1506
|
if self.usage is not None:
|
|
@@ -1495,6 +1511,14 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult(TeaMo
|
|
|
1495
1511
|
m = m or dict()
|
|
1496
1512
|
if m.get('generateFinished') is not None:
|
|
1497
1513
|
self.generate_finished = m.get('generateFinished')
|
|
1514
|
+
if m.get('index') is not None:
|
|
1515
|
+
self.index = m.get('index')
|
|
1516
|
+
if m.get('modelId') is not None:
|
|
1517
|
+
self.model_id = m.get('modelId')
|
|
1518
|
+
if m.get('modelReduce') is not None:
|
|
1519
|
+
self.model_reduce = m.get('modelReduce')
|
|
1520
|
+
if m.get('reasonText') is not None:
|
|
1521
|
+
self.reason_text = m.get('reasonText')
|
|
1498
1522
|
if m.get('text') is not None:
|
|
1499
1523
|
self.text = m.get('text')
|
|
1500
1524
|
if m.get('usage') is not None:
|
|
@@ -1503,6 +1527,104 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult(TeaMo
|
|
|
1503
1527
|
return self
|
|
1504
1528
|
|
|
1505
1529
|
|
|
1530
|
+
class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResultsUsage(TeaModel):
|
|
1531
|
+
def __init__(
|
|
1532
|
+
self,
|
|
1533
|
+
input_tokens: int = None,
|
|
1534
|
+
output_tokens: int = None,
|
|
1535
|
+
total_tokens: int = None,
|
|
1536
|
+
):
|
|
1537
|
+
self.input_tokens = input_tokens
|
|
1538
|
+
self.output_tokens = output_tokens
|
|
1539
|
+
self.total_tokens = total_tokens
|
|
1540
|
+
|
|
1541
|
+
def validate(self):
|
|
1542
|
+
pass
|
|
1543
|
+
|
|
1544
|
+
def to_map(self):
|
|
1545
|
+
_map = super().to_map()
|
|
1546
|
+
if _map is not None:
|
|
1547
|
+
return _map
|
|
1548
|
+
|
|
1549
|
+
result = dict()
|
|
1550
|
+
if self.input_tokens is not None:
|
|
1551
|
+
result['inputTokens'] = self.input_tokens
|
|
1552
|
+
if self.output_tokens is not None:
|
|
1553
|
+
result['outputTokens'] = self.output_tokens
|
|
1554
|
+
if self.total_tokens is not None:
|
|
1555
|
+
result['totalTokens'] = self.total_tokens
|
|
1556
|
+
return result
|
|
1557
|
+
|
|
1558
|
+
def from_map(self, m: dict = None):
|
|
1559
|
+
m = m or dict()
|
|
1560
|
+
if m.get('inputTokens') is not None:
|
|
1561
|
+
self.input_tokens = m.get('inputTokens')
|
|
1562
|
+
if m.get('outputTokens') is not None:
|
|
1563
|
+
self.output_tokens = m.get('outputTokens')
|
|
1564
|
+
if m.get('totalTokens') is not None:
|
|
1565
|
+
self.total_tokens = m.get('totalTokens')
|
|
1566
|
+
return self
|
|
1567
|
+
|
|
1568
|
+
|
|
1569
|
+
class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResults(TeaModel):
|
|
1570
|
+
def __init__(
|
|
1571
|
+
self,
|
|
1572
|
+
generate_finished: bool = None,
|
|
1573
|
+
index: int = None,
|
|
1574
|
+
model_id: str = None,
|
|
1575
|
+
reason_text: str = None,
|
|
1576
|
+
text: str = None,
|
|
1577
|
+
usage: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResultsUsage = None,
|
|
1578
|
+
):
|
|
1579
|
+
self.generate_finished = generate_finished
|
|
1580
|
+
self.index = index
|
|
1581
|
+
self.model_id = model_id
|
|
1582
|
+
self.reason_text = reason_text
|
|
1583
|
+
self.text = text
|
|
1584
|
+
self.usage = usage
|
|
1585
|
+
|
|
1586
|
+
def validate(self):
|
|
1587
|
+
if self.usage:
|
|
1588
|
+
self.usage.validate()
|
|
1589
|
+
|
|
1590
|
+
def to_map(self):
|
|
1591
|
+
_map = super().to_map()
|
|
1592
|
+
if _map is not None:
|
|
1593
|
+
return _map
|
|
1594
|
+
|
|
1595
|
+
result = dict()
|
|
1596
|
+
if self.generate_finished is not None:
|
|
1597
|
+
result['generateFinished'] = self.generate_finished
|
|
1598
|
+
if self.index is not None:
|
|
1599
|
+
result['index'] = self.index
|
|
1600
|
+
if self.model_id is not None:
|
|
1601
|
+
result['modelId'] = self.model_id
|
|
1602
|
+
if self.reason_text is not None:
|
|
1603
|
+
result['reasonText'] = self.reason_text
|
|
1604
|
+
if self.text is not None:
|
|
1605
|
+
result['text'] = self.text
|
|
1606
|
+
if self.usage is not None:
|
|
1607
|
+
result['usage'] = self.usage.to_map()
|
|
1608
|
+
return result
|
|
1609
|
+
|
|
1610
|
+
def from_map(self, m: dict = None):
|
|
1611
|
+
m = m or dict()
|
|
1612
|
+
if m.get('generateFinished') is not None:
|
|
1613
|
+
self.generate_finished = m.get('generateFinished')
|
|
1614
|
+
if m.get('index') is not None:
|
|
1615
|
+
self.index = m.get('index')
|
|
1616
|
+
if m.get('modelId') is not None:
|
|
1617
|
+
self.model_id = m.get('modelId')
|
|
1618
|
+
if m.get('reasonText') is not None:
|
|
1619
|
+
self.reason_text = m.get('reasonText')
|
|
1620
|
+
if m.get('text') is not None:
|
|
1621
|
+
self.text = m.get('text')
|
|
1622
|
+
if m.get('usage') is not None:
|
|
1623
|
+
temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResultsUsage()
|
|
1624
|
+
self.usage = temp_model.from_map(m['usage'])
|
|
1625
|
+
return self
|
|
1626
|
+
|
|
1627
|
+
|
|
1506
1628
|
class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoMindMappingGenerateResultUsage(TeaModel):
|
|
1507
1629
|
def __init__(
|
|
1508
1630
|
self,
|
|
@@ -1794,6 +1916,7 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
|
|
|
1794
1916
|
video_analysis_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoAnalysisResult = None,
|
|
1795
1917
|
video_caption_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoCaptionResult = None,
|
|
1796
1918
|
video_generate_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult = None,
|
|
1919
|
+
video_generate_results: List[GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResults] = None,
|
|
1797
1920
|
video_mind_mapping_generate_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoMindMappingGenerateResult = None,
|
|
1798
1921
|
video_title_generate_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoTitleGenerateResult = None,
|
|
1799
1922
|
):
|
|
@@ -1801,6 +1924,7 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
|
|
|
1801
1924
|
self.video_analysis_result = video_analysis_result
|
|
1802
1925
|
self.video_caption_result = video_caption_result
|
|
1803
1926
|
self.video_generate_result = video_generate_result
|
|
1927
|
+
self.video_generate_results = video_generate_results
|
|
1804
1928
|
self.video_mind_mapping_generate_result = video_mind_mapping_generate_result
|
|
1805
1929
|
self.video_title_generate_result = video_title_generate_result
|
|
1806
1930
|
|
|
@@ -1811,6 +1935,10 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
|
|
|
1811
1935
|
self.video_caption_result.validate()
|
|
1812
1936
|
if self.video_generate_result:
|
|
1813
1937
|
self.video_generate_result.validate()
|
|
1938
|
+
if self.video_generate_results:
|
|
1939
|
+
for k in self.video_generate_results:
|
|
1940
|
+
if k:
|
|
1941
|
+
k.validate()
|
|
1814
1942
|
if self.video_mind_mapping_generate_result:
|
|
1815
1943
|
self.video_mind_mapping_generate_result.validate()
|
|
1816
1944
|
if self.video_title_generate_result:
|
|
@@ -1830,6 +1958,10 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
|
|
|
1830
1958
|
result['videoCaptionResult'] = self.video_caption_result.to_map()
|
|
1831
1959
|
if self.video_generate_result is not None:
|
|
1832
1960
|
result['videoGenerateResult'] = self.video_generate_result.to_map()
|
|
1961
|
+
result['videoGenerateResults'] = []
|
|
1962
|
+
if self.video_generate_results is not None:
|
|
1963
|
+
for k in self.video_generate_results:
|
|
1964
|
+
result['videoGenerateResults'].append(k.to_map() if k else None)
|
|
1833
1965
|
if self.video_mind_mapping_generate_result is not None:
|
|
1834
1966
|
result['videoMindMappingGenerateResult'] = self.video_mind_mapping_generate_result.to_map()
|
|
1835
1967
|
if self.video_title_generate_result is not None:
|
|
@@ -1849,6 +1981,11 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
|
|
|
1849
1981
|
if m.get('videoGenerateResult') is not None:
|
|
1850
1982
|
temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult()
|
|
1851
1983
|
self.video_generate_result = temp_model.from_map(m['videoGenerateResult'])
|
|
1984
|
+
self.video_generate_results = []
|
|
1985
|
+
if m.get('videoGenerateResults') is not None:
|
|
1986
|
+
for k in m.get('videoGenerateResults'):
|
|
1987
|
+
temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResults()
|
|
1988
|
+
self.video_generate_results.append(temp_model.from_map(k))
|
|
1852
1989
|
if m.get('videoMindMappingGenerateResult') is not None:
|
|
1853
1990
|
temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoMindMappingGenerateResult()
|
|
1854
1991
|
self.video_mind_mapping_generate_result = temp_model.from_map(m['videoMindMappingGenerateResult'])
|
|
@@ -3430,7 +3567,6 @@ class RunHotTopicSummaryRequestStepForCustomSummaryStyleConfig(TeaModel):
|
|
|
3430
3567
|
summary_prompt: str = None,
|
|
3431
3568
|
):
|
|
3432
3569
|
self.summary_image_count = summary_image_count
|
|
3433
|
-
# This parameter is required.
|
|
3434
3570
|
self.summary_model = summary_model
|
|
3435
3571
|
self.summary_prompt = summary_prompt
|
|
3436
3572
|
|
|
@@ -6916,6 +7052,45 @@ class RunVideoAnalysisRequestFrameSampleMethod(TeaModel):
|
|
|
6916
7052
|
return self
|
|
6917
7053
|
|
|
6918
7054
|
|
|
7055
|
+
class RunVideoAnalysisRequestTextProcessTasks(TeaModel):
|
|
7056
|
+
def __init__(
|
|
7057
|
+
self,
|
|
7058
|
+
model_custom_prompt_template: str = None,
|
|
7059
|
+
model_custom_prompt_template_id: str = None,
|
|
7060
|
+
model_id: str = None,
|
|
7061
|
+
):
|
|
7062
|
+
self.model_custom_prompt_template = model_custom_prompt_template
|
|
7063
|
+
self.model_custom_prompt_template_id = model_custom_prompt_template_id
|
|
7064
|
+
self.model_id = model_id
|
|
7065
|
+
|
|
7066
|
+
def validate(self):
|
|
7067
|
+
pass
|
|
7068
|
+
|
|
7069
|
+
def to_map(self):
|
|
7070
|
+
_map = super().to_map()
|
|
7071
|
+
if _map is not None:
|
|
7072
|
+
return _map
|
|
7073
|
+
|
|
7074
|
+
result = dict()
|
|
7075
|
+
if self.model_custom_prompt_template is not None:
|
|
7076
|
+
result['modelCustomPromptTemplate'] = self.model_custom_prompt_template
|
|
7077
|
+
if self.model_custom_prompt_template_id is not None:
|
|
7078
|
+
result['modelCustomPromptTemplateId'] = self.model_custom_prompt_template_id
|
|
7079
|
+
if self.model_id is not None:
|
|
7080
|
+
result['modelId'] = self.model_id
|
|
7081
|
+
return result
|
|
7082
|
+
|
|
7083
|
+
def from_map(self, m: dict = None):
|
|
7084
|
+
m = m or dict()
|
|
7085
|
+
if m.get('modelCustomPromptTemplate') is not None:
|
|
7086
|
+
self.model_custom_prompt_template = m.get('modelCustomPromptTemplate')
|
|
7087
|
+
if m.get('modelCustomPromptTemplateId') is not None:
|
|
7088
|
+
self.model_custom_prompt_template_id = m.get('modelCustomPromptTemplateId')
|
|
7089
|
+
if m.get('modelId') is not None:
|
|
7090
|
+
self.model_id = m.get('modelId')
|
|
7091
|
+
return self
|
|
7092
|
+
|
|
7093
|
+
|
|
6919
7094
|
class RunVideoAnalysisRequestVideoRoles(TeaModel):
|
|
6920
7095
|
def __init__(
|
|
6921
7096
|
self,
|
|
@@ -6958,6 +7133,7 @@ class RunVideoAnalysisRequestVideoRoles(TeaModel):
|
|
|
6958
7133
|
class RunVideoAnalysisRequest(TeaModel):
|
|
6959
7134
|
def __init__(
|
|
6960
7135
|
self,
|
|
7136
|
+
face_identity_similarity_min_score: float = None,
|
|
6961
7137
|
frame_sample_method: RunVideoAnalysisRequestFrameSampleMethod = None,
|
|
6962
7138
|
generate_options: List[str] = None,
|
|
6963
7139
|
language: str = None,
|
|
@@ -6967,12 +7143,15 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
6967
7143
|
original_session_id: str = None,
|
|
6968
7144
|
snapshot_interval: float = None,
|
|
6969
7145
|
task_id: str = None,
|
|
7146
|
+
text_process_tasks: List[RunVideoAnalysisRequestTextProcessTasks] = None,
|
|
6970
7147
|
video_extra_info: str = None,
|
|
6971
7148
|
video_model_custom_prompt_template: str = None,
|
|
6972
7149
|
video_model_id: str = None,
|
|
6973
7150
|
video_roles: List[RunVideoAnalysisRequestVideoRoles] = None,
|
|
7151
|
+
video_shot_face_identity_count: int = None,
|
|
6974
7152
|
video_url: str = None,
|
|
6975
7153
|
):
|
|
7154
|
+
self.face_identity_similarity_min_score = face_identity_similarity_min_score
|
|
6976
7155
|
self.frame_sample_method = frame_sample_method
|
|
6977
7156
|
self.generate_options = generate_options
|
|
6978
7157
|
self.language = language
|
|
@@ -6982,15 +7161,21 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
6982
7161
|
self.original_session_id = original_session_id
|
|
6983
7162
|
self.snapshot_interval = snapshot_interval
|
|
6984
7163
|
self.task_id = task_id
|
|
7164
|
+
self.text_process_tasks = text_process_tasks
|
|
6985
7165
|
self.video_extra_info = video_extra_info
|
|
6986
7166
|
self.video_model_custom_prompt_template = video_model_custom_prompt_template
|
|
6987
7167
|
self.video_model_id = video_model_id
|
|
6988
7168
|
self.video_roles = video_roles
|
|
7169
|
+
self.video_shot_face_identity_count = video_shot_face_identity_count
|
|
6989
7170
|
self.video_url = video_url
|
|
6990
7171
|
|
|
6991
7172
|
def validate(self):
|
|
6992
7173
|
if self.frame_sample_method:
|
|
6993
7174
|
self.frame_sample_method.validate()
|
|
7175
|
+
if self.text_process_tasks:
|
|
7176
|
+
for k in self.text_process_tasks:
|
|
7177
|
+
if k:
|
|
7178
|
+
k.validate()
|
|
6994
7179
|
if self.video_roles:
|
|
6995
7180
|
for k in self.video_roles:
|
|
6996
7181
|
if k:
|
|
@@ -7002,6 +7187,8 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
7002
7187
|
return _map
|
|
7003
7188
|
|
|
7004
7189
|
result = dict()
|
|
7190
|
+
if self.face_identity_similarity_min_score is not None:
|
|
7191
|
+
result['faceIdentitySimilarityMinScore'] = self.face_identity_similarity_min_score
|
|
7005
7192
|
if self.frame_sample_method is not None:
|
|
7006
7193
|
result['frameSampleMethod'] = self.frame_sample_method.to_map()
|
|
7007
7194
|
if self.generate_options is not None:
|
|
@@ -7020,6 +7207,10 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
7020
7207
|
result['snapshotInterval'] = self.snapshot_interval
|
|
7021
7208
|
if self.task_id is not None:
|
|
7022
7209
|
result['taskId'] = self.task_id
|
|
7210
|
+
result['textProcessTasks'] = []
|
|
7211
|
+
if self.text_process_tasks is not None:
|
|
7212
|
+
for k in self.text_process_tasks:
|
|
7213
|
+
result['textProcessTasks'].append(k.to_map() if k else None)
|
|
7023
7214
|
if self.video_extra_info is not None:
|
|
7024
7215
|
result['videoExtraInfo'] = self.video_extra_info
|
|
7025
7216
|
if self.video_model_custom_prompt_template is not None:
|
|
@@ -7030,12 +7221,16 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
7030
7221
|
if self.video_roles is not None:
|
|
7031
7222
|
for k in self.video_roles:
|
|
7032
7223
|
result['videoRoles'].append(k.to_map() if k else None)
|
|
7224
|
+
if self.video_shot_face_identity_count is not None:
|
|
7225
|
+
result['videoShotFaceIdentityCount'] = self.video_shot_face_identity_count
|
|
7033
7226
|
if self.video_url is not None:
|
|
7034
7227
|
result['videoUrl'] = self.video_url
|
|
7035
7228
|
return result
|
|
7036
7229
|
|
|
7037
7230
|
def from_map(self, m: dict = None):
|
|
7038
7231
|
m = m or dict()
|
|
7232
|
+
if m.get('faceIdentitySimilarityMinScore') is not None:
|
|
7233
|
+
self.face_identity_similarity_min_score = m.get('faceIdentitySimilarityMinScore')
|
|
7039
7234
|
if m.get('frameSampleMethod') is not None:
|
|
7040
7235
|
temp_model = RunVideoAnalysisRequestFrameSampleMethod()
|
|
7041
7236
|
self.frame_sample_method = temp_model.from_map(m['frameSampleMethod'])
|
|
@@ -7055,6 +7250,11 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
7055
7250
|
self.snapshot_interval = m.get('snapshotInterval')
|
|
7056
7251
|
if m.get('taskId') is not None:
|
|
7057
7252
|
self.task_id = m.get('taskId')
|
|
7253
|
+
self.text_process_tasks = []
|
|
7254
|
+
if m.get('textProcessTasks') is not None:
|
|
7255
|
+
for k in m.get('textProcessTasks'):
|
|
7256
|
+
temp_model = RunVideoAnalysisRequestTextProcessTasks()
|
|
7257
|
+
self.text_process_tasks.append(temp_model.from_map(k))
|
|
7058
7258
|
if m.get('videoExtraInfo') is not None:
|
|
7059
7259
|
self.video_extra_info = m.get('videoExtraInfo')
|
|
7060
7260
|
if m.get('videoModelCustomPromptTemplate') is not None:
|
|
@@ -7066,6 +7266,8 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
7066
7266
|
for k in m.get('videoRoles'):
|
|
7067
7267
|
temp_model = RunVideoAnalysisRequestVideoRoles()
|
|
7068
7268
|
self.video_roles.append(temp_model.from_map(k))
|
|
7269
|
+
if m.get('videoShotFaceIdentityCount') is not None:
|
|
7270
|
+
self.video_shot_face_identity_count = m.get('videoShotFaceIdentityCount')
|
|
7069
7271
|
if m.get('videoUrl') is not None:
|
|
7070
7272
|
self.video_url = m.get('videoUrl')
|
|
7071
7273
|
return self
|
|
@@ -7074,6 +7276,7 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
7074
7276
|
class RunVideoAnalysisShrinkRequest(TeaModel):
|
|
7075
7277
|
def __init__(
|
|
7076
7278
|
self,
|
|
7279
|
+
face_identity_similarity_min_score: float = None,
|
|
7077
7280
|
frame_sample_method_shrink: str = None,
|
|
7078
7281
|
generate_options_shrink: str = None,
|
|
7079
7282
|
language: str = None,
|
|
@@ -7083,12 +7286,15 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
|
|
|
7083
7286
|
original_session_id: str = None,
|
|
7084
7287
|
snapshot_interval: float = None,
|
|
7085
7288
|
task_id: str = None,
|
|
7289
|
+
text_process_tasks_shrink: str = None,
|
|
7086
7290
|
video_extra_info: str = None,
|
|
7087
7291
|
video_model_custom_prompt_template: str = None,
|
|
7088
7292
|
video_model_id: str = None,
|
|
7089
7293
|
video_roles_shrink: str = None,
|
|
7294
|
+
video_shot_face_identity_count: int = None,
|
|
7090
7295
|
video_url: str = None,
|
|
7091
7296
|
):
|
|
7297
|
+
self.face_identity_similarity_min_score = face_identity_similarity_min_score
|
|
7092
7298
|
self.frame_sample_method_shrink = frame_sample_method_shrink
|
|
7093
7299
|
self.generate_options_shrink = generate_options_shrink
|
|
7094
7300
|
self.language = language
|
|
@@ -7098,10 +7304,12 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
|
|
|
7098
7304
|
self.original_session_id = original_session_id
|
|
7099
7305
|
self.snapshot_interval = snapshot_interval
|
|
7100
7306
|
self.task_id = task_id
|
|
7307
|
+
self.text_process_tasks_shrink = text_process_tasks_shrink
|
|
7101
7308
|
self.video_extra_info = video_extra_info
|
|
7102
7309
|
self.video_model_custom_prompt_template = video_model_custom_prompt_template
|
|
7103
7310
|
self.video_model_id = video_model_id
|
|
7104
7311
|
self.video_roles_shrink = video_roles_shrink
|
|
7312
|
+
self.video_shot_face_identity_count = video_shot_face_identity_count
|
|
7105
7313
|
self.video_url = video_url
|
|
7106
7314
|
|
|
7107
7315
|
def validate(self):
|
|
@@ -7113,6 +7321,8 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
|
|
|
7113
7321
|
return _map
|
|
7114
7322
|
|
|
7115
7323
|
result = dict()
|
|
7324
|
+
if self.face_identity_similarity_min_score is not None:
|
|
7325
|
+
result['faceIdentitySimilarityMinScore'] = self.face_identity_similarity_min_score
|
|
7116
7326
|
if self.frame_sample_method_shrink is not None:
|
|
7117
7327
|
result['frameSampleMethod'] = self.frame_sample_method_shrink
|
|
7118
7328
|
if self.generate_options_shrink is not None:
|
|
@@ -7131,6 +7341,8 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
|
|
|
7131
7341
|
result['snapshotInterval'] = self.snapshot_interval
|
|
7132
7342
|
if self.task_id is not None:
|
|
7133
7343
|
result['taskId'] = self.task_id
|
|
7344
|
+
if self.text_process_tasks_shrink is not None:
|
|
7345
|
+
result['textProcessTasks'] = self.text_process_tasks_shrink
|
|
7134
7346
|
if self.video_extra_info is not None:
|
|
7135
7347
|
result['videoExtraInfo'] = self.video_extra_info
|
|
7136
7348
|
if self.video_model_custom_prompt_template is not None:
|
|
@@ -7139,12 +7351,16 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
|
|
|
7139
7351
|
result['videoModelId'] = self.video_model_id
|
|
7140
7352
|
if self.video_roles_shrink is not None:
|
|
7141
7353
|
result['videoRoles'] = self.video_roles_shrink
|
|
7354
|
+
if self.video_shot_face_identity_count is not None:
|
|
7355
|
+
result['videoShotFaceIdentityCount'] = self.video_shot_face_identity_count
|
|
7142
7356
|
if self.video_url is not None:
|
|
7143
7357
|
result['videoUrl'] = self.video_url
|
|
7144
7358
|
return result
|
|
7145
7359
|
|
|
7146
7360
|
def from_map(self, m: dict = None):
|
|
7147
7361
|
m = m or dict()
|
|
7362
|
+
if m.get('faceIdentitySimilarityMinScore') is not None:
|
|
7363
|
+
self.face_identity_similarity_min_score = m.get('faceIdentitySimilarityMinScore')
|
|
7148
7364
|
if m.get('frameSampleMethod') is not None:
|
|
7149
7365
|
self.frame_sample_method_shrink = m.get('frameSampleMethod')
|
|
7150
7366
|
if m.get('generateOptions') is not None:
|
|
@@ -7163,6 +7379,8 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
|
|
|
7163
7379
|
self.snapshot_interval = m.get('snapshotInterval')
|
|
7164
7380
|
if m.get('taskId') is not None:
|
|
7165
7381
|
self.task_id = m.get('taskId')
|
|
7382
|
+
if m.get('textProcessTasks') is not None:
|
|
7383
|
+
self.text_process_tasks_shrink = m.get('textProcessTasks')
|
|
7166
7384
|
if m.get('videoExtraInfo') is not None:
|
|
7167
7385
|
self.video_extra_info = m.get('videoExtraInfo')
|
|
7168
7386
|
if m.get('videoModelCustomPromptTemplate') is not None:
|
|
@@ -7171,6 +7389,8 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
|
|
|
7171
7389
|
self.video_model_id = m.get('videoModelId')
|
|
7172
7390
|
if m.get('videoRoles') is not None:
|
|
7173
7391
|
self.video_roles_shrink = m.get('videoRoles')
|
|
7392
|
+
if m.get('videoShotFaceIdentityCount') is not None:
|
|
7393
|
+
self.video_shot_face_identity_count = m.get('videoShotFaceIdentityCount')
|
|
7174
7394
|
if m.get('videoUrl') is not None:
|
|
7175
7395
|
self.video_url = m.get('videoUrl')
|
|
7176
7396
|
return self
|
|
@@ -7514,14 +7734,18 @@ class RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult(TeaModel):
|
|
|
7514
7734
|
def __init__(
|
|
7515
7735
|
self,
|
|
7516
7736
|
generate_finished: bool = None,
|
|
7737
|
+
index: int = None,
|
|
7517
7738
|
model_id: str = None,
|
|
7518
7739
|
model_reduce: bool = None,
|
|
7740
|
+
reason_text: str = None,
|
|
7519
7741
|
text: str = None,
|
|
7520
7742
|
usage: RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResultUsage = None,
|
|
7521
7743
|
):
|
|
7522
7744
|
self.generate_finished = generate_finished
|
|
7745
|
+
self.index = index
|
|
7523
7746
|
self.model_id = model_id
|
|
7524
7747
|
self.model_reduce = model_reduce
|
|
7748
|
+
self.reason_text = reason_text
|
|
7525
7749
|
self.text = text
|
|
7526
7750
|
self.usage = usage
|
|
7527
7751
|
|
|
@@ -7537,10 +7761,14 @@ class RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult(TeaModel):
|
|
|
7537
7761
|
result = dict()
|
|
7538
7762
|
if self.generate_finished is not None:
|
|
7539
7763
|
result['generateFinished'] = self.generate_finished
|
|
7764
|
+
if self.index is not None:
|
|
7765
|
+
result['index'] = self.index
|
|
7540
7766
|
if self.model_id is not None:
|
|
7541
7767
|
result['modelId'] = self.model_id
|
|
7542
7768
|
if self.model_reduce is not None:
|
|
7543
7769
|
result['modelReduce'] = self.model_reduce
|
|
7770
|
+
if self.reason_text is not None:
|
|
7771
|
+
result['reasonText'] = self.reason_text
|
|
7544
7772
|
if self.text is not None:
|
|
7545
7773
|
result['text'] = self.text
|
|
7546
7774
|
if self.usage is not None:
|
|
@@ -7551,10 +7779,14 @@ class RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult(TeaModel):
|
|
|
7551
7779
|
m = m or dict()
|
|
7552
7780
|
if m.get('generateFinished') is not None:
|
|
7553
7781
|
self.generate_finished = m.get('generateFinished')
|
|
7782
|
+
if m.get('index') is not None:
|
|
7783
|
+
self.index = m.get('index')
|
|
7554
7784
|
if m.get('modelId') is not None:
|
|
7555
7785
|
self.model_id = m.get('modelId')
|
|
7556
7786
|
if m.get('modelReduce') is not None:
|
|
7557
7787
|
self.model_reduce = m.get('modelReduce')
|
|
7788
|
+
if m.get('reasonText') is not None:
|
|
7789
|
+
self.reason_text = m.get('reasonText')
|
|
7558
7790
|
if m.get('text') is not None:
|
|
7559
7791
|
self.text = m.get('text')
|
|
7560
7792
|
if m.get('usage') is not None:
|
|
@@ -7563,6 +7795,104 @@ class RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult(TeaModel):
|
|
|
7563
7795
|
return self
|
|
7564
7796
|
|
|
7565
7797
|
|
|
7798
|
+
class RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResultsUsage(TeaModel):
|
|
7799
|
+
def __init__(
|
|
7800
|
+
self,
|
|
7801
|
+
input_tokens: int = None,
|
|
7802
|
+
output_tokens: int = None,
|
|
7803
|
+
total_tokens: int = None,
|
|
7804
|
+
):
|
|
7805
|
+
self.input_tokens = input_tokens
|
|
7806
|
+
self.output_tokens = output_tokens
|
|
7807
|
+
self.total_tokens = total_tokens
|
|
7808
|
+
|
|
7809
|
+
def validate(self):
|
|
7810
|
+
pass
|
|
7811
|
+
|
|
7812
|
+
def to_map(self):
|
|
7813
|
+
_map = super().to_map()
|
|
7814
|
+
if _map is not None:
|
|
7815
|
+
return _map
|
|
7816
|
+
|
|
7817
|
+
result = dict()
|
|
7818
|
+
if self.input_tokens is not None:
|
|
7819
|
+
result['inputTokens'] = self.input_tokens
|
|
7820
|
+
if self.output_tokens is not None:
|
|
7821
|
+
result['outputTokens'] = self.output_tokens
|
|
7822
|
+
if self.total_tokens is not None:
|
|
7823
|
+
result['totalTokens'] = self.total_tokens
|
|
7824
|
+
return result
|
|
7825
|
+
|
|
7826
|
+
def from_map(self, m: dict = None):
|
|
7827
|
+
m = m or dict()
|
|
7828
|
+
if m.get('inputTokens') is not None:
|
|
7829
|
+
self.input_tokens = m.get('inputTokens')
|
|
7830
|
+
if m.get('outputTokens') is not None:
|
|
7831
|
+
self.output_tokens = m.get('outputTokens')
|
|
7832
|
+
if m.get('totalTokens') is not None:
|
|
7833
|
+
self.total_tokens = m.get('totalTokens')
|
|
7834
|
+
return self
|
|
7835
|
+
|
|
7836
|
+
|
|
7837
|
+
class RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResults(TeaModel):
|
|
7838
|
+
def __init__(
|
|
7839
|
+
self,
|
|
7840
|
+
generate_finished: bool = None,
|
|
7841
|
+
index: int = None,
|
|
7842
|
+
model_id: str = None,
|
|
7843
|
+
reason_text: str = None,
|
|
7844
|
+
text: str = None,
|
|
7845
|
+
usage: RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResultsUsage = None,
|
|
7846
|
+
):
|
|
7847
|
+
self.generate_finished = generate_finished
|
|
7848
|
+
self.index = index
|
|
7849
|
+
self.model_id = model_id
|
|
7850
|
+
self.reason_text = reason_text
|
|
7851
|
+
self.text = text
|
|
7852
|
+
self.usage = usage
|
|
7853
|
+
|
|
7854
|
+
def validate(self):
|
|
7855
|
+
if self.usage:
|
|
7856
|
+
self.usage.validate()
|
|
7857
|
+
|
|
7858
|
+
def to_map(self):
|
|
7859
|
+
_map = super().to_map()
|
|
7860
|
+
if _map is not None:
|
|
7861
|
+
return _map
|
|
7862
|
+
|
|
7863
|
+
result = dict()
|
|
7864
|
+
if self.generate_finished is not None:
|
|
7865
|
+
result['generateFinished'] = self.generate_finished
|
|
7866
|
+
if self.index is not None:
|
|
7867
|
+
result['index'] = self.index
|
|
7868
|
+
if self.model_id is not None:
|
|
7869
|
+
result['modelId'] = self.model_id
|
|
7870
|
+
if self.reason_text is not None:
|
|
7871
|
+
result['reasonText'] = self.reason_text
|
|
7872
|
+
if self.text is not None:
|
|
7873
|
+
result['text'] = self.text
|
|
7874
|
+
if self.usage is not None:
|
|
7875
|
+
result['usage'] = self.usage.to_map()
|
|
7876
|
+
return result
|
|
7877
|
+
|
|
7878
|
+
def from_map(self, m: dict = None):
|
|
7879
|
+
m = m or dict()
|
|
7880
|
+
if m.get('generateFinished') is not None:
|
|
7881
|
+
self.generate_finished = m.get('generateFinished')
|
|
7882
|
+
if m.get('index') is not None:
|
|
7883
|
+
self.index = m.get('index')
|
|
7884
|
+
if m.get('modelId') is not None:
|
|
7885
|
+
self.model_id = m.get('modelId')
|
|
7886
|
+
if m.get('reasonText') is not None:
|
|
7887
|
+
self.reason_text = m.get('reasonText')
|
|
7888
|
+
if m.get('text') is not None:
|
|
7889
|
+
self.text = m.get('text')
|
|
7890
|
+
if m.get('usage') is not None:
|
|
7891
|
+
temp_model = RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResultsUsage()
|
|
7892
|
+
self.usage = temp_model.from_map(m['usage'])
|
|
7893
|
+
return self
|
|
7894
|
+
|
|
7895
|
+
|
|
7566
7896
|
class RunVideoAnalysisResponseBodyPayloadOutputVideoMindMappingGenerateResultUsage(TeaModel):
|
|
7567
7897
|
def __init__(
|
|
7568
7898
|
self,
|
|
@@ -7999,6 +8329,7 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
|
|
|
7999
8329
|
video_analysis_result: RunVideoAnalysisResponseBodyPayloadOutputVideoAnalysisResult = None,
|
|
8000
8330
|
video_caption_result: RunVideoAnalysisResponseBodyPayloadOutputVideoCaptionResult = None,
|
|
8001
8331
|
video_generate_result: RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult = None,
|
|
8332
|
+
video_generate_results: List[RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResults] = None,
|
|
8002
8333
|
video_mind_mapping_generate_result: RunVideoAnalysisResponseBodyPayloadOutputVideoMindMappingGenerateResult = None,
|
|
8003
8334
|
video_shot_snapshot_result: RunVideoAnalysisResponseBodyPayloadOutputVideoShotSnapshotResult = None,
|
|
8004
8335
|
video_title_generate_result: RunVideoAnalysisResponseBodyPayloadOutputVideoTitleGenerateResult = None,
|
|
@@ -8007,6 +8338,7 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
|
|
|
8007
8338
|
self.video_analysis_result = video_analysis_result
|
|
8008
8339
|
self.video_caption_result = video_caption_result
|
|
8009
8340
|
self.video_generate_result = video_generate_result
|
|
8341
|
+
self.video_generate_results = video_generate_results
|
|
8010
8342
|
self.video_mind_mapping_generate_result = video_mind_mapping_generate_result
|
|
8011
8343
|
self.video_shot_snapshot_result = video_shot_snapshot_result
|
|
8012
8344
|
self.video_title_generate_result = video_title_generate_result
|
|
@@ -8018,6 +8350,10 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
|
|
|
8018
8350
|
self.video_caption_result.validate()
|
|
8019
8351
|
if self.video_generate_result:
|
|
8020
8352
|
self.video_generate_result.validate()
|
|
8353
|
+
if self.video_generate_results:
|
|
8354
|
+
for k in self.video_generate_results:
|
|
8355
|
+
if k:
|
|
8356
|
+
k.validate()
|
|
8021
8357
|
if self.video_mind_mapping_generate_result:
|
|
8022
8358
|
self.video_mind_mapping_generate_result.validate()
|
|
8023
8359
|
if self.video_shot_snapshot_result:
|
|
@@ -8039,6 +8375,10 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
|
|
|
8039
8375
|
result['videoCaptionResult'] = self.video_caption_result.to_map()
|
|
8040
8376
|
if self.video_generate_result is not None:
|
|
8041
8377
|
result['videoGenerateResult'] = self.video_generate_result.to_map()
|
|
8378
|
+
result['videoGenerateResults'] = []
|
|
8379
|
+
if self.video_generate_results is not None:
|
|
8380
|
+
for k in self.video_generate_results:
|
|
8381
|
+
result['videoGenerateResults'].append(k.to_map() if k else None)
|
|
8042
8382
|
if self.video_mind_mapping_generate_result is not None:
|
|
8043
8383
|
result['videoMindMappingGenerateResult'] = self.video_mind_mapping_generate_result.to_map()
|
|
8044
8384
|
if self.video_shot_snapshot_result is not None:
|
|
@@ -8060,6 +8400,11 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
|
|
|
8060
8400
|
if m.get('videoGenerateResult') is not None:
|
|
8061
8401
|
temp_model = RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult()
|
|
8062
8402
|
self.video_generate_result = temp_model.from_map(m['videoGenerateResult'])
|
|
8403
|
+
self.video_generate_results = []
|
|
8404
|
+
if m.get('videoGenerateResults') is not None:
|
|
8405
|
+
for k in m.get('videoGenerateResults'):
|
|
8406
|
+
temp_model = RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResults()
|
|
8407
|
+
self.video_generate_results.append(temp_model.from_map(k))
|
|
8063
8408
|
if m.get('videoMindMappingGenerateResult') is not None:
|
|
8064
8409
|
temp_model = RunVideoAnalysisResponseBodyPayloadOutputVideoMindMappingGenerateResult()
|
|
8065
8410
|
self.video_mind_mapping_generate_result = temp_model.from_map(m['videoMindMappingGenerateResult'])
|
|
@@ -8580,6 +8925,45 @@ class SubmitVideoAnalysisTaskRequestFrameSampleMethod(TeaModel):
|
|
|
8580
8925
|
return self
|
|
8581
8926
|
|
|
8582
8927
|
|
|
8928
|
+
class SubmitVideoAnalysisTaskRequestTextProcessTasks(TeaModel):
|
|
8929
|
+
def __init__(
|
|
8930
|
+
self,
|
|
8931
|
+
model_custom_prompt_template: str = None,
|
|
8932
|
+
model_custom_prompt_template_id: str = None,
|
|
8933
|
+
model_id: str = None,
|
|
8934
|
+
):
|
|
8935
|
+
self.model_custom_prompt_template = model_custom_prompt_template
|
|
8936
|
+
self.model_custom_prompt_template_id = model_custom_prompt_template_id
|
|
8937
|
+
self.model_id = model_id
|
|
8938
|
+
|
|
8939
|
+
def validate(self):
|
|
8940
|
+
pass
|
|
8941
|
+
|
|
8942
|
+
def to_map(self):
|
|
8943
|
+
_map = super().to_map()
|
|
8944
|
+
if _map is not None:
|
|
8945
|
+
return _map
|
|
8946
|
+
|
|
8947
|
+
result = dict()
|
|
8948
|
+
if self.model_custom_prompt_template is not None:
|
|
8949
|
+
result['modelCustomPromptTemplate'] = self.model_custom_prompt_template
|
|
8950
|
+
if self.model_custom_prompt_template_id is not None:
|
|
8951
|
+
result['modelCustomPromptTemplateId'] = self.model_custom_prompt_template_id
|
|
8952
|
+
if self.model_id is not None:
|
|
8953
|
+
result['modelId'] = self.model_id
|
|
8954
|
+
return result
|
|
8955
|
+
|
|
8956
|
+
def from_map(self, m: dict = None):
|
|
8957
|
+
m = m or dict()
|
|
8958
|
+
if m.get('modelCustomPromptTemplate') is not None:
|
|
8959
|
+
self.model_custom_prompt_template = m.get('modelCustomPromptTemplate')
|
|
8960
|
+
if m.get('modelCustomPromptTemplateId') is not None:
|
|
8961
|
+
self.model_custom_prompt_template_id = m.get('modelCustomPromptTemplateId')
|
|
8962
|
+
if m.get('modelId') is not None:
|
|
8963
|
+
self.model_id = m.get('modelId')
|
|
8964
|
+
return self
|
|
8965
|
+
|
|
8966
|
+
|
|
8583
8967
|
class SubmitVideoAnalysisTaskRequestVideoRoles(TeaModel):
|
|
8584
8968
|
def __init__(
|
|
8585
8969
|
self,
|
|
@@ -8622,6 +9006,7 @@ class SubmitVideoAnalysisTaskRequestVideoRoles(TeaModel):
|
|
|
8622
9006
|
class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
8623
9007
|
def __init__(
|
|
8624
9008
|
self,
|
|
9009
|
+
face_identity_similarity_min_score: float = None,
|
|
8625
9010
|
frame_sample_method: SubmitVideoAnalysisTaskRequestFrameSampleMethod = None,
|
|
8626
9011
|
generate_options: List[str] = None,
|
|
8627
9012
|
language: str = None,
|
|
@@ -8629,12 +9014,15 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
8629
9014
|
model_custom_prompt_template_id: str = None,
|
|
8630
9015
|
model_id: str = None,
|
|
8631
9016
|
snapshot_interval: float = None,
|
|
9017
|
+
text_process_tasks: List[SubmitVideoAnalysisTaskRequestTextProcessTasks] = None,
|
|
8632
9018
|
video_extra_info: str = None,
|
|
8633
9019
|
video_model_custom_prompt_template: str = None,
|
|
8634
9020
|
video_model_id: str = None,
|
|
8635
9021
|
video_roles: List[SubmitVideoAnalysisTaskRequestVideoRoles] = None,
|
|
9022
|
+
video_shot_face_identity_count: int = None,
|
|
8636
9023
|
video_url: str = None,
|
|
8637
9024
|
):
|
|
9025
|
+
self.face_identity_similarity_min_score = face_identity_similarity_min_score
|
|
8638
9026
|
self.frame_sample_method = frame_sample_method
|
|
8639
9027
|
self.generate_options = generate_options
|
|
8640
9028
|
self.language = language
|
|
@@ -8642,16 +9030,22 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
8642
9030
|
self.model_custom_prompt_template_id = model_custom_prompt_template_id
|
|
8643
9031
|
self.model_id = model_id
|
|
8644
9032
|
self.snapshot_interval = snapshot_interval
|
|
9033
|
+
self.text_process_tasks = text_process_tasks
|
|
8645
9034
|
self.video_extra_info = video_extra_info
|
|
8646
9035
|
self.video_model_custom_prompt_template = video_model_custom_prompt_template
|
|
8647
9036
|
self.video_model_id = video_model_id
|
|
8648
9037
|
self.video_roles = video_roles
|
|
9038
|
+
self.video_shot_face_identity_count = video_shot_face_identity_count
|
|
8649
9039
|
# This parameter is required.
|
|
8650
9040
|
self.video_url = video_url
|
|
8651
9041
|
|
|
8652
9042
|
def validate(self):
|
|
8653
9043
|
if self.frame_sample_method:
|
|
8654
9044
|
self.frame_sample_method.validate()
|
|
9045
|
+
if self.text_process_tasks:
|
|
9046
|
+
for k in self.text_process_tasks:
|
|
9047
|
+
if k:
|
|
9048
|
+
k.validate()
|
|
8655
9049
|
if self.video_roles:
|
|
8656
9050
|
for k in self.video_roles:
|
|
8657
9051
|
if k:
|
|
@@ -8663,6 +9057,8 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
8663
9057
|
return _map
|
|
8664
9058
|
|
|
8665
9059
|
result = dict()
|
|
9060
|
+
if self.face_identity_similarity_min_score is not None:
|
|
9061
|
+
result['faceIdentitySimilarityMinScore'] = self.face_identity_similarity_min_score
|
|
8666
9062
|
if self.frame_sample_method is not None:
|
|
8667
9063
|
result['frameSampleMethod'] = self.frame_sample_method.to_map()
|
|
8668
9064
|
if self.generate_options is not None:
|
|
@@ -8677,6 +9073,10 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
8677
9073
|
result['modelId'] = self.model_id
|
|
8678
9074
|
if self.snapshot_interval is not None:
|
|
8679
9075
|
result['snapshotInterval'] = self.snapshot_interval
|
|
9076
|
+
result['textProcessTasks'] = []
|
|
9077
|
+
if self.text_process_tasks is not None:
|
|
9078
|
+
for k in self.text_process_tasks:
|
|
9079
|
+
result['textProcessTasks'].append(k.to_map() if k else None)
|
|
8680
9080
|
if self.video_extra_info is not None:
|
|
8681
9081
|
result['videoExtraInfo'] = self.video_extra_info
|
|
8682
9082
|
if self.video_model_custom_prompt_template is not None:
|
|
@@ -8687,12 +9087,16 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
8687
9087
|
if self.video_roles is not None:
|
|
8688
9088
|
for k in self.video_roles:
|
|
8689
9089
|
result['videoRoles'].append(k.to_map() if k else None)
|
|
9090
|
+
if self.video_shot_face_identity_count is not None:
|
|
9091
|
+
result['videoShotFaceIdentityCount'] = self.video_shot_face_identity_count
|
|
8690
9092
|
if self.video_url is not None:
|
|
8691
9093
|
result['videoUrl'] = self.video_url
|
|
8692
9094
|
return result
|
|
8693
9095
|
|
|
8694
9096
|
def from_map(self, m: dict = None):
|
|
8695
9097
|
m = m or dict()
|
|
9098
|
+
if m.get('faceIdentitySimilarityMinScore') is not None:
|
|
9099
|
+
self.face_identity_similarity_min_score = m.get('faceIdentitySimilarityMinScore')
|
|
8696
9100
|
if m.get('frameSampleMethod') is not None:
|
|
8697
9101
|
temp_model = SubmitVideoAnalysisTaskRequestFrameSampleMethod()
|
|
8698
9102
|
self.frame_sample_method = temp_model.from_map(m['frameSampleMethod'])
|
|
@@ -8708,6 +9112,11 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
8708
9112
|
self.model_id = m.get('modelId')
|
|
8709
9113
|
if m.get('snapshotInterval') is not None:
|
|
8710
9114
|
self.snapshot_interval = m.get('snapshotInterval')
|
|
9115
|
+
self.text_process_tasks = []
|
|
9116
|
+
if m.get('textProcessTasks') is not None:
|
|
9117
|
+
for k in m.get('textProcessTasks'):
|
|
9118
|
+
temp_model = SubmitVideoAnalysisTaskRequestTextProcessTasks()
|
|
9119
|
+
self.text_process_tasks.append(temp_model.from_map(k))
|
|
8711
9120
|
if m.get('videoExtraInfo') is not None:
|
|
8712
9121
|
self.video_extra_info = m.get('videoExtraInfo')
|
|
8713
9122
|
if m.get('videoModelCustomPromptTemplate') is not None:
|
|
@@ -8719,6 +9128,8 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
8719
9128
|
for k in m.get('videoRoles'):
|
|
8720
9129
|
temp_model = SubmitVideoAnalysisTaskRequestVideoRoles()
|
|
8721
9130
|
self.video_roles.append(temp_model.from_map(k))
|
|
9131
|
+
if m.get('videoShotFaceIdentityCount') is not None:
|
|
9132
|
+
self.video_shot_face_identity_count = m.get('videoShotFaceIdentityCount')
|
|
8722
9133
|
if m.get('videoUrl') is not None:
|
|
8723
9134
|
self.video_url = m.get('videoUrl')
|
|
8724
9135
|
return self
|
|
@@ -8727,6 +9138,7 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
8727
9138
|
class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
|
|
8728
9139
|
def __init__(
|
|
8729
9140
|
self,
|
|
9141
|
+
face_identity_similarity_min_score: float = None,
|
|
8730
9142
|
frame_sample_method_shrink: str = None,
|
|
8731
9143
|
generate_options_shrink: str = None,
|
|
8732
9144
|
language: str = None,
|
|
@@ -8734,12 +9146,15 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
|
|
|
8734
9146
|
model_custom_prompt_template_id: str = None,
|
|
8735
9147
|
model_id: str = None,
|
|
8736
9148
|
snapshot_interval: float = None,
|
|
9149
|
+
text_process_tasks_shrink: str = None,
|
|
8737
9150
|
video_extra_info: str = None,
|
|
8738
9151
|
video_model_custom_prompt_template: str = None,
|
|
8739
9152
|
video_model_id: str = None,
|
|
8740
9153
|
video_roles_shrink: str = None,
|
|
9154
|
+
video_shot_face_identity_count: int = None,
|
|
8741
9155
|
video_url: str = None,
|
|
8742
9156
|
):
|
|
9157
|
+
self.face_identity_similarity_min_score = face_identity_similarity_min_score
|
|
8743
9158
|
self.frame_sample_method_shrink = frame_sample_method_shrink
|
|
8744
9159
|
self.generate_options_shrink = generate_options_shrink
|
|
8745
9160
|
self.language = language
|
|
@@ -8747,10 +9162,12 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
|
|
|
8747
9162
|
self.model_custom_prompt_template_id = model_custom_prompt_template_id
|
|
8748
9163
|
self.model_id = model_id
|
|
8749
9164
|
self.snapshot_interval = snapshot_interval
|
|
9165
|
+
self.text_process_tasks_shrink = text_process_tasks_shrink
|
|
8750
9166
|
self.video_extra_info = video_extra_info
|
|
8751
9167
|
self.video_model_custom_prompt_template = video_model_custom_prompt_template
|
|
8752
9168
|
self.video_model_id = video_model_id
|
|
8753
9169
|
self.video_roles_shrink = video_roles_shrink
|
|
9170
|
+
self.video_shot_face_identity_count = video_shot_face_identity_count
|
|
8754
9171
|
# This parameter is required.
|
|
8755
9172
|
self.video_url = video_url
|
|
8756
9173
|
|
|
@@ -8763,6 +9180,8 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
|
|
|
8763
9180
|
return _map
|
|
8764
9181
|
|
|
8765
9182
|
result = dict()
|
|
9183
|
+
if self.face_identity_similarity_min_score is not None:
|
|
9184
|
+
result['faceIdentitySimilarityMinScore'] = self.face_identity_similarity_min_score
|
|
8766
9185
|
if self.frame_sample_method_shrink is not None:
|
|
8767
9186
|
result['frameSampleMethod'] = self.frame_sample_method_shrink
|
|
8768
9187
|
if self.generate_options_shrink is not None:
|
|
@@ -8777,6 +9196,8 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
|
|
|
8777
9196
|
result['modelId'] = self.model_id
|
|
8778
9197
|
if self.snapshot_interval is not None:
|
|
8779
9198
|
result['snapshotInterval'] = self.snapshot_interval
|
|
9199
|
+
if self.text_process_tasks_shrink is not None:
|
|
9200
|
+
result['textProcessTasks'] = self.text_process_tasks_shrink
|
|
8780
9201
|
if self.video_extra_info is not None:
|
|
8781
9202
|
result['videoExtraInfo'] = self.video_extra_info
|
|
8782
9203
|
if self.video_model_custom_prompt_template is not None:
|
|
@@ -8785,12 +9206,16 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
|
|
|
8785
9206
|
result['videoModelId'] = self.video_model_id
|
|
8786
9207
|
if self.video_roles_shrink is not None:
|
|
8787
9208
|
result['videoRoles'] = self.video_roles_shrink
|
|
9209
|
+
if self.video_shot_face_identity_count is not None:
|
|
9210
|
+
result['videoShotFaceIdentityCount'] = self.video_shot_face_identity_count
|
|
8788
9211
|
if self.video_url is not None:
|
|
8789
9212
|
result['videoUrl'] = self.video_url
|
|
8790
9213
|
return result
|
|
8791
9214
|
|
|
8792
9215
|
def from_map(self, m: dict = None):
|
|
8793
9216
|
m = m or dict()
|
|
9217
|
+
if m.get('faceIdentitySimilarityMinScore') is not None:
|
|
9218
|
+
self.face_identity_similarity_min_score = m.get('faceIdentitySimilarityMinScore')
|
|
8794
9219
|
if m.get('frameSampleMethod') is not None:
|
|
8795
9220
|
self.frame_sample_method_shrink = m.get('frameSampleMethod')
|
|
8796
9221
|
if m.get('generateOptions') is not None:
|
|
@@ -8805,6 +9230,8 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
|
|
|
8805
9230
|
self.model_id = m.get('modelId')
|
|
8806
9231
|
if m.get('snapshotInterval') is not None:
|
|
8807
9232
|
self.snapshot_interval = m.get('snapshotInterval')
|
|
9233
|
+
if m.get('textProcessTasks') is not None:
|
|
9234
|
+
self.text_process_tasks_shrink = m.get('textProcessTasks')
|
|
8808
9235
|
if m.get('videoExtraInfo') is not None:
|
|
8809
9236
|
self.video_extra_info = m.get('videoExtraInfo')
|
|
8810
9237
|
if m.get('videoModelCustomPromptTemplate') is not None:
|
|
@@ -8813,6 +9240,8 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
|
|
|
8813
9240
|
self.video_model_id = m.get('videoModelId')
|
|
8814
9241
|
if m.get('videoRoles') is not None:
|
|
8815
9242
|
self.video_roles_shrink = m.get('videoRoles')
|
|
9243
|
+
if m.get('videoShotFaceIdentityCount') is not None:
|
|
9244
|
+
self.video_shot_face_identity_count = m.get('videoShotFaceIdentityCount')
|
|
8816
9245
|
if m.get('videoUrl') is not None:
|
|
8817
9246
|
self.video_url = m.get('videoUrl')
|
|
8818
9247
|
return self
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.1
|
|
2
2
|
Name: alibabacloud-quanmiaolightapp20240801
|
|
3
|
-
Version: 2.6.
|
|
3
|
+
Version: 2.6.2
|
|
4
4
|
Summary: Alibaba Cloud QuanMiaoLightApp (20240801) SDK Library for Python
|
|
5
5
|
Home-page: https://github.com/aliyun/alibabacloud-python-sdk
|
|
6
6
|
Author: Alibaba Cloud SDK
|
alibabacloud_quanmiaolightapp20240801-2.6.0/alibabacloud_quanmiaolightapp20240801/__init__.py
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
__version__ = '2.6.0'
|
{alibabacloud_quanmiaolightapp20240801-2.6.0 → alibabacloud_quanmiaolightapp20240801-2.6.2}/LICENSE
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|