alibabacloud-quanmiaolightapp20240801 2.13.0__tar.gz → 2.13.2__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/ChangeLog.md +12 -0
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/PKG-INFO +1 -1
- alibabacloud_quanmiaolightapp20240801-2.13.2/alibabacloud_quanmiaolightapp20240801/__init__.py +1 -0
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/alibabacloud_quanmiaolightapp20240801/client.py +16 -0
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/alibabacloud_quanmiaolightapp20240801/models.py +548 -0
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/alibabacloud_quanmiaolightapp20240801.egg-info/PKG-INFO +1 -1
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/setup.py +1 -1
- alibabacloud_quanmiaolightapp20240801-2.13.0/alibabacloud_quanmiaolightapp20240801/__init__.py +0 -1
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/LICENSE +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/MANIFEST.in +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/README-CN.md +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/README.md +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/alibabacloud_quanmiaolightapp20240801.egg-info/SOURCES.txt +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/alibabacloud_quanmiaolightapp20240801.egg-info/dependency_links.txt +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/alibabacloud_quanmiaolightapp20240801.egg-info/requires.txt +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/alibabacloud_quanmiaolightapp20240801.egg-info/top_level.txt +0 -0
- {alibabacloud_quanmiaolightapp20240801-2.13.0 → alibabacloud_quanmiaolightapp20240801-2.13.2}/setup.cfg +0 -0
|
@@ -1,3 +1,15 @@
|
|
|
1
|
+
2025-10-21 Version: 2.13.1
|
|
2
|
+
- Generated python 2024-08-01 for QuanMiaoLightApp.
|
|
3
|
+
|
|
4
|
+
2025-10-21 Version: 2.13.0
|
|
5
|
+
- Support API GetVideoDetectShotConfig.
|
|
6
|
+
- Support API GetVideoDetectShotTask.
|
|
7
|
+
- Support API RunVideoDetectShot.
|
|
8
|
+
- Support API SubmitVideoDetectShotTask.
|
|
9
|
+
- Support API UpdateVideoDetectShotConfig.
|
|
10
|
+
- Support API UpdateVideoDetectShotTask.
|
|
11
|
+
|
|
12
|
+
|
|
1
13
|
2025-10-17 Version: 2.12.1
|
|
2
14
|
- Update API RunVideoAnalysis: add request parameters splitType.
|
|
3
15
|
- Update API SubmitVideoAnalysisTask: add request parameters splitType.
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.1
|
|
2
2
|
Name: alibabacloud_quanmiaolightapp20240801
|
|
3
|
-
Version: 2.13.
|
|
3
|
+
Version: 2.13.2
|
|
4
4
|
Summary: Alibaba Cloud QuanMiaoLightApp (20240801) SDK Library for Python
|
|
5
5
|
Home-page: https://github.com/aliyun/alibabacloud-python-sdk
|
|
6
6
|
Author: Alibaba Cloud SDK
|
alibabacloud_quanmiaolightapp20240801-2.13.2/alibabacloud_quanmiaolightapp20240801/__init__.py
ADDED
|
@@ -0,0 +1 @@
|
|
|
1
|
+
__version__ = '2.13.2'
|
|
@@ -3571,6 +3571,8 @@ class Client(OpenApiClient):
|
|
|
3571
3571
|
UtilClient.validate_model(tmp_req)
|
|
3572
3572
|
request = quan_miao_light_app_20240801_models.RunVideoAnalysisShrinkRequest()
|
|
3573
3573
|
OpenApiUtilClient.convert(tmp_req, request)
|
|
3574
|
+
if not UtilClient.is_unset(tmp_req.add_document_param):
|
|
3575
|
+
request.add_document_param_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.add_document_param, 'addDocumentParam', 'json')
|
|
3574
3576
|
if not UtilClient.is_unset(tmp_req.exclude_generate_options):
|
|
3575
3577
|
request.exclude_generate_options_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.exclude_generate_options, 'excludeGenerateOptions', 'json')
|
|
3576
3578
|
if not UtilClient.is_unset(tmp_req.frame_sample_method):
|
|
@@ -3584,6 +3586,8 @@ class Client(OpenApiClient):
|
|
|
3584
3586
|
if not UtilClient.is_unset(tmp_req.video_roles):
|
|
3585
3587
|
request.video_roles_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.video_roles, 'videoRoles', 'json')
|
|
3586
3588
|
body = {}
|
|
3589
|
+
if not UtilClient.is_unset(request.add_document_param_shrink):
|
|
3590
|
+
body['addDocumentParam'] = request.add_document_param_shrink
|
|
3587
3591
|
if not UtilClient.is_unset(request.auto_role_recognition_video_url):
|
|
3588
3592
|
body['autoRoleRecognitionVideoUrl'] = request.auto_role_recognition_video_url
|
|
3589
3593
|
if not UtilClient.is_unset(request.exclude_generate_options_shrink):
|
|
@@ -3666,6 +3670,8 @@ class Client(OpenApiClient):
|
|
|
3666
3670
|
UtilClient.validate_model(tmp_req)
|
|
3667
3671
|
request = quan_miao_light_app_20240801_models.RunVideoAnalysisShrinkRequest()
|
|
3668
3672
|
OpenApiUtilClient.convert(tmp_req, request)
|
|
3673
|
+
if not UtilClient.is_unset(tmp_req.add_document_param):
|
|
3674
|
+
request.add_document_param_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.add_document_param, 'addDocumentParam', 'json')
|
|
3669
3675
|
if not UtilClient.is_unset(tmp_req.exclude_generate_options):
|
|
3670
3676
|
request.exclude_generate_options_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.exclude_generate_options, 'excludeGenerateOptions', 'json')
|
|
3671
3677
|
if not UtilClient.is_unset(tmp_req.frame_sample_method):
|
|
@@ -3679,6 +3685,8 @@ class Client(OpenApiClient):
|
|
|
3679
3685
|
if not UtilClient.is_unset(tmp_req.video_roles):
|
|
3680
3686
|
request.video_roles_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.video_roles, 'videoRoles', 'json')
|
|
3681
3687
|
body = {}
|
|
3688
|
+
if not UtilClient.is_unset(request.add_document_param_shrink):
|
|
3689
|
+
body['addDocumentParam'] = request.add_document_param_shrink
|
|
3682
3690
|
if not UtilClient.is_unset(request.auto_role_recognition_video_url):
|
|
3683
3691
|
body['autoRoleRecognitionVideoUrl'] = request.auto_role_recognition_video_url
|
|
3684
3692
|
if not UtilClient.is_unset(request.exclude_generate_options_shrink):
|
|
@@ -4419,6 +4427,8 @@ class Client(OpenApiClient):
|
|
|
4419
4427
|
UtilClient.validate_model(tmp_req)
|
|
4420
4428
|
request = quan_miao_light_app_20240801_models.SubmitVideoAnalysisTaskShrinkRequest()
|
|
4421
4429
|
OpenApiUtilClient.convert(tmp_req, request)
|
|
4430
|
+
if not UtilClient.is_unset(tmp_req.add_document_param):
|
|
4431
|
+
request.add_document_param_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.add_document_param, 'addDocumentParam', 'json')
|
|
4422
4432
|
if not UtilClient.is_unset(tmp_req.exclude_generate_options):
|
|
4423
4433
|
request.exclude_generate_options_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.exclude_generate_options, 'excludeGenerateOptions', 'json')
|
|
4424
4434
|
if not UtilClient.is_unset(tmp_req.frame_sample_method):
|
|
@@ -4432,6 +4442,8 @@ class Client(OpenApiClient):
|
|
|
4432
4442
|
if not UtilClient.is_unset(tmp_req.video_roles):
|
|
4433
4443
|
request.video_roles_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.video_roles, 'videoRoles', 'json')
|
|
4434
4444
|
body = {}
|
|
4445
|
+
if not UtilClient.is_unset(request.add_document_param_shrink):
|
|
4446
|
+
body['addDocumentParam'] = request.add_document_param_shrink
|
|
4435
4447
|
if not UtilClient.is_unset(request.auto_role_recognition_video_url):
|
|
4436
4448
|
body['autoRoleRecognitionVideoUrl'] = request.auto_role_recognition_video_url
|
|
4437
4449
|
if not UtilClient.is_unset(request.deduplication_id):
|
|
@@ -4512,6 +4524,8 @@ class Client(OpenApiClient):
|
|
|
4512
4524
|
UtilClient.validate_model(tmp_req)
|
|
4513
4525
|
request = quan_miao_light_app_20240801_models.SubmitVideoAnalysisTaskShrinkRequest()
|
|
4514
4526
|
OpenApiUtilClient.convert(tmp_req, request)
|
|
4527
|
+
if not UtilClient.is_unset(tmp_req.add_document_param):
|
|
4528
|
+
request.add_document_param_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.add_document_param, 'addDocumentParam', 'json')
|
|
4515
4529
|
if not UtilClient.is_unset(tmp_req.exclude_generate_options):
|
|
4516
4530
|
request.exclude_generate_options_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.exclude_generate_options, 'excludeGenerateOptions', 'json')
|
|
4517
4531
|
if not UtilClient.is_unset(tmp_req.frame_sample_method):
|
|
@@ -4525,6 +4539,8 @@ class Client(OpenApiClient):
|
|
|
4525
4539
|
if not UtilClient.is_unset(tmp_req.video_roles):
|
|
4526
4540
|
request.video_roles_shrink = OpenApiUtilClient.array_to_string_with_specified_style(tmp_req.video_roles, 'videoRoles', 'json')
|
|
4527
4541
|
body = {}
|
|
4542
|
+
if not UtilClient.is_unset(request.add_document_param_shrink):
|
|
4543
|
+
body['addDocumentParam'] = request.add_document_param_shrink
|
|
4528
4544
|
if not UtilClient.is_unset(request.auto_role_recognition_video_url):
|
|
4529
4545
|
body['autoRoleRecognitionVideoUrl'] = request.auto_role_recognition_video_url
|
|
4530
4546
|
if not UtilClient.is_unset(request.deduplication_id):
|
|
@@ -2269,13 +2269,66 @@ class GetVideoAnalysisTaskResponseBodyDataHeader(TeaModel):
|
|
|
2269
2269
|
return self
|
|
2270
2270
|
|
|
2271
2271
|
|
|
2272
|
+
class GetVideoAnalysisTaskResponseBodyDataPayloadOutputAddDatasetDocumentsResult(TeaModel):
|
|
2273
|
+
def __init__(
|
|
2274
|
+
self,
|
|
2275
|
+
doc_id: str = None,
|
|
2276
|
+
doc_uuid: str = None,
|
|
2277
|
+
error_message: str = None,
|
|
2278
|
+
status: int = None,
|
|
2279
|
+
title: str = None,
|
|
2280
|
+
):
|
|
2281
|
+
self.doc_id = doc_id
|
|
2282
|
+
self.doc_uuid = doc_uuid
|
|
2283
|
+
self.error_message = error_message
|
|
2284
|
+
self.status = status
|
|
2285
|
+
self.title = title
|
|
2286
|
+
|
|
2287
|
+
def validate(self):
|
|
2288
|
+
pass
|
|
2289
|
+
|
|
2290
|
+
def to_map(self):
|
|
2291
|
+
_map = super().to_map()
|
|
2292
|
+
if _map is not None:
|
|
2293
|
+
return _map
|
|
2294
|
+
|
|
2295
|
+
result = dict()
|
|
2296
|
+
if self.doc_id is not None:
|
|
2297
|
+
result['docId'] = self.doc_id
|
|
2298
|
+
if self.doc_uuid is not None:
|
|
2299
|
+
result['docUuid'] = self.doc_uuid
|
|
2300
|
+
if self.error_message is not None:
|
|
2301
|
+
result['errorMessage'] = self.error_message
|
|
2302
|
+
if self.status is not None:
|
|
2303
|
+
result['status'] = self.status
|
|
2304
|
+
if self.title is not None:
|
|
2305
|
+
result['title'] = self.title
|
|
2306
|
+
return result
|
|
2307
|
+
|
|
2308
|
+
def from_map(self, m: dict = None):
|
|
2309
|
+
m = m or dict()
|
|
2310
|
+
if m.get('docId') is not None:
|
|
2311
|
+
self.doc_id = m.get('docId')
|
|
2312
|
+
if m.get('docUuid') is not None:
|
|
2313
|
+
self.doc_uuid = m.get('docUuid')
|
|
2314
|
+
if m.get('errorMessage') is not None:
|
|
2315
|
+
self.error_message = m.get('errorMessage')
|
|
2316
|
+
if m.get('status') is not None:
|
|
2317
|
+
self.status = m.get('status')
|
|
2318
|
+
if m.get('title') is not None:
|
|
2319
|
+
self.title = m.get('title')
|
|
2320
|
+
return self
|
|
2321
|
+
|
|
2322
|
+
|
|
2272
2323
|
class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoAnalysisResultUsage(TeaModel):
|
|
2273
2324
|
def __init__(
|
|
2274
2325
|
self,
|
|
2326
|
+
image_tokens: int = None,
|
|
2275
2327
|
input_tokens: int = None,
|
|
2276
2328
|
output_tokens: int = None,
|
|
2277
2329
|
total_tokens: int = None,
|
|
2278
2330
|
):
|
|
2331
|
+
self.image_tokens = image_tokens
|
|
2279
2332
|
self.input_tokens = input_tokens
|
|
2280
2333
|
self.output_tokens = output_tokens
|
|
2281
2334
|
self.total_tokens = total_tokens
|
|
@@ -2289,6 +2342,8 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoAnalysisResultUsage(
|
|
|
2289
2342
|
return _map
|
|
2290
2343
|
|
|
2291
2344
|
result = dict()
|
|
2345
|
+
if self.image_tokens is not None:
|
|
2346
|
+
result['imageTokens'] = self.image_tokens
|
|
2292
2347
|
if self.input_tokens is not None:
|
|
2293
2348
|
result['inputTokens'] = self.input_tokens
|
|
2294
2349
|
if self.output_tokens is not None:
|
|
@@ -2299,6 +2354,8 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoAnalysisResultUsage(
|
|
|
2299
2354
|
|
|
2300
2355
|
def from_map(self, m: dict = None):
|
|
2301
2356
|
m = m or dict()
|
|
2357
|
+
if m.get('imageTokens') is not None:
|
|
2358
|
+
self.image_tokens = m.get('imageTokens')
|
|
2302
2359
|
if m.get('inputTokens') is not None:
|
|
2303
2360
|
self.input_tokens = m.get('inputTokens')
|
|
2304
2361
|
if m.get('outputTokens') is not None:
|
|
@@ -2403,6 +2460,116 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoAnalysisResult(TeaMo
|
|
|
2403
2460
|
return self
|
|
2404
2461
|
|
|
2405
2462
|
|
|
2463
|
+
class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoCalculatorResultItems(TeaModel):
|
|
2464
|
+
def __init__(
|
|
2465
|
+
self,
|
|
2466
|
+
input_expense: float = None,
|
|
2467
|
+
input_token: int = None,
|
|
2468
|
+
name: str = None,
|
|
2469
|
+
output_expense: float = None,
|
|
2470
|
+
output_token: int = None,
|
|
2471
|
+
time: int = None,
|
|
2472
|
+
time_expense: float = None,
|
|
2473
|
+
total_expense: float = None,
|
|
2474
|
+
type: str = None,
|
|
2475
|
+
):
|
|
2476
|
+
self.input_expense = input_expense
|
|
2477
|
+
self.input_token = input_token
|
|
2478
|
+
self.name = name
|
|
2479
|
+
self.output_expense = output_expense
|
|
2480
|
+
self.output_token = output_token
|
|
2481
|
+
self.time = time
|
|
2482
|
+
self.time_expense = time_expense
|
|
2483
|
+
self.total_expense = total_expense
|
|
2484
|
+
self.type = type
|
|
2485
|
+
|
|
2486
|
+
def validate(self):
|
|
2487
|
+
pass
|
|
2488
|
+
|
|
2489
|
+
def to_map(self):
|
|
2490
|
+
_map = super().to_map()
|
|
2491
|
+
if _map is not None:
|
|
2492
|
+
return _map
|
|
2493
|
+
|
|
2494
|
+
result = dict()
|
|
2495
|
+
if self.input_expense is not None:
|
|
2496
|
+
result['inputExpense'] = self.input_expense
|
|
2497
|
+
if self.input_token is not None:
|
|
2498
|
+
result['inputToken'] = self.input_token
|
|
2499
|
+
if self.name is not None:
|
|
2500
|
+
result['name'] = self.name
|
|
2501
|
+
if self.output_expense is not None:
|
|
2502
|
+
result['outputExpense'] = self.output_expense
|
|
2503
|
+
if self.output_token is not None:
|
|
2504
|
+
result['outputToken'] = self.output_token
|
|
2505
|
+
if self.time is not None:
|
|
2506
|
+
result['time'] = self.time
|
|
2507
|
+
if self.time_expense is not None:
|
|
2508
|
+
result['timeExpense'] = self.time_expense
|
|
2509
|
+
if self.total_expense is not None:
|
|
2510
|
+
result['totalExpense'] = self.total_expense
|
|
2511
|
+
if self.type is not None:
|
|
2512
|
+
result['type'] = self.type
|
|
2513
|
+
return result
|
|
2514
|
+
|
|
2515
|
+
def from_map(self, m: dict = None):
|
|
2516
|
+
m = m or dict()
|
|
2517
|
+
if m.get('inputExpense') is not None:
|
|
2518
|
+
self.input_expense = m.get('inputExpense')
|
|
2519
|
+
if m.get('inputToken') is not None:
|
|
2520
|
+
self.input_token = m.get('inputToken')
|
|
2521
|
+
if m.get('name') is not None:
|
|
2522
|
+
self.name = m.get('name')
|
|
2523
|
+
if m.get('outputExpense') is not None:
|
|
2524
|
+
self.output_expense = m.get('outputExpense')
|
|
2525
|
+
if m.get('outputToken') is not None:
|
|
2526
|
+
self.output_token = m.get('outputToken')
|
|
2527
|
+
if m.get('time') is not None:
|
|
2528
|
+
self.time = m.get('time')
|
|
2529
|
+
if m.get('timeExpense') is not None:
|
|
2530
|
+
self.time_expense = m.get('timeExpense')
|
|
2531
|
+
if m.get('totalExpense') is not None:
|
|
2532
|
+
self.total_expense = m.get('totalExpense')
|
|
2533
|
+
if m.get('type') is not None:
|
|
2534
|
+
self.type = m.get('type')
|
|
2535
|
+
return self
|
|
2536
|
+
|
|
2537
|
+
|
|
2538
|
+
class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoCalculatorResult(TeaModel):
|
|
2539
|
+
def __init__(
|
|
2540
|
+
self,
|
|
2541
|
+
items: List[GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoCalculatorResultItems] = None,
|
|
2542
|
+
):
|
|
2543
|
+
self.items = items
|
|
2544
|
+
|
|
2545
|
+
def validate(self):
|
|
2546
|
+
if self.items:
|
|
2547
|
+
for k in self.items:
|
|
2548
|
+
if k:
|
|
2549
|
+
k.validate()
|
|
2550
|
+
|
|
2551
|
+
def to_map(self):
|
|
2552
|
+
_map = super().to_map()
|
|
2553
|
+
if _map is not None:
|
|
2554
|
+
return _map
|
|
2555
|
+
|
|
2556
|
+
result = dict()
|
|
2557
|
+
result['items'] = []
|
|
2558
|
+
if self.items is not None:
|
|
2559
|
+
for k in self.items:
|
|
2560
|
+
result['items'].append(k.to_map() if k else None)
|
|
2561
|
+
return result
|
|
2562
|
+
|
|
2563
|
+
def from_map(self, m: dict = None):
|
|
2564
|
+
m = m or dict()
|
|
2565
|
+
self.items = []
|
|
2566
|
+
if m.get('items') is not None:
|
|
2567
|
+
for k in m.get('items'):
|
|
2568
|
+
temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoCalculatorResultItems()
|
|
2569
|
+
self.items.append(temp_model.from_map(k))
|
|
2570
|
+
return self
|
|
2571
|
+
|
|
2572
|
+
|
|
2406
2573
|
class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoCaptionResultVideoCaptions(TeaModel):
|
|
2407
2574
|
def __init__(
|
|
2408
2575
|
self,
|
|
@@ -3129,8 +3296,10 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoTitleGenerateResult(
|
|
|
3129
3296
|
class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
|
|
3130
3297
|
def __init__(
|
|
3131
3298
|
self,
|
|
3299
|
+
add_dataset_documents_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputAddDatasetDocumentsResult = None,
|
|
3132
3300
|
result_json_file_url: str = None,
|
|
3133
3301
|
video_analysis_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoAnalysisResult = None,
|
|
3302
|
+
video_calculator_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoCalculatorResult = None,
|
|
3134
3303
|
video_caption_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoCaptionResult = None,
|
|
3135
3304
|
video_generate_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResult = None,
|
|
3136
3305
|
video_generate_results: List[GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoGenerateResults] = None,
|
|
@@ -3138,8 +3307,10 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
|
|
|
3138
3307
|
video_role_recognition_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoRoleRecognitionResult = None,
|
|
3139
3308
|
video_title_generate_result: GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoTitleGenerateResult = None,
|
|
3140
3309
|
):
|
|
3310
|
+
self.add_dataset_documents_result = add_dataset_documents_result
|
|
3141
3311
|
self.result_json_file_url = result_json_file_url
|
|
3142
3312
|
self.video_analysis_result = video_analysis_result
|
|
3313
|
+
self.video_calculator_result = video_calculator_result
|
|
3143
3314
|
self.video_caption_result = video_caption_result
|
|
3144
3315
|
self.video_generate_result = video_generate_result
|
|
3145
3316
|
self.video_generate_results = video_generate_results
|
|
@@ -3148,8 +3319,12 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
|
|
|
3148
3319
|
self.video_title_generate_result = video_title_generate_result
|
|
3149
3320
|
|
|
3150
3321
|
def validate(self):
|
|
3322
|
+
if self.add_dataset_documents_result:
|
|
3323
|
+
self.add_dataset_documents_result.validate()
|
|
3151
3324
|
if self.video_analysis_result:
|
|
3152
3325
|
self.video_analysis_result.validate()
|
|
3326
|
+
if self.video_calculator_result:
|
|
3327
|
+
self.video_calculator_result.validate()
|
|
3153
3328
|
if self.video_caption_result:
|
|
3154
3329
|
self.video_caption_result.validate()
|
|
3155
3330
|
if self.video_generate_result:
|
|
@@ -3171,10 +3346,14 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
|
|
|
3171
3346
|
return _map
|
|
3172
3347
|
|
|
3173
3348
|
result = dict()
|
|
3349
|
+
if self.add_dataset_documents_result is not None:
|
|
3350
|
+
result['addDatasetDocumentsResult'] = self.add_dataset_documents_result.to_map()
|
|
3174
3351
|
if self.result_json_file_url is not None:
|
|
3175
3352
|
result['resultJsonFileUrl'] = self.result_json_file_url
|
|
3176
3353
|
if self.video_analysis_result is not None:
|
|
3177
3354
|
result['videoAnalysisResult'] = self.video_analysis_result.to_map()
|
|
3355
|
+
if self.video_calculator_result is not None:
|
|
3356
|
+
result['videoCalculatorResult'] = self.video_calculator_result.to_map()
|
|
3178
3357
|
if self.video_caption_result is not None:
|
|
3179
3358
|
result['videoCaptionResult'] = self.video_caption_result.to_map()
|
|
3180
3359
|
if self.video_generate_result is not None:
|
|
@@ -3193,11 +3372,17 @@ class GetVideoAnalysisTaskResponseBodyDataPayloadOutput(TeaModel):
|
|
|
3193
3372
|
|
|
3194
3373
|
def from_map(self, m: dict = None):
|
|
3195
3374
|
m = m or dict()
|
|
3375
|
+
if m.get('addDatasetDocumentsResult') is not None:
|
|
3376
|
+
temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputAddDatasetDocumentsResult()
|
|
3377
|
+
self.add_dataset_documents_result = temp_model.from_map(m['addDatasetDocumentsResult'])
|
|
3196
3378
|
if m.get('resultJsonFileUrl') is not None:
|
|
3197
3379
|
self.result_json_file_url = m.get('resultJsonFileUrl')
|
|
3198
3380
|
if m.get('videoAnalysisResult') is not None:
|
|
3199
3381
|
temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoAnalysisResult()
|
|
3200
3382
|
self.video_analysis_result = temp_model.from_map(m['videoAnalysisResult'])
|
|
3383
|
+
if m.get('videoCalculatorResult') is not None:
|
|
3384
|
+
temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoCalculatorResult()
|
|
3385
|
+
self.video_calculator_result = temp_model.from_map(m['videoCalculatorResult'])
|
|
3201
3386
|
if m.get('videoCaptionResult') is not None:
|
|
3202
3387
|
temp_model = GetVideoAnalysisTaskResponseBodyDataPayloadOutputVideoCaptionResult()
|
|
3203
3388
|
self.video_caption_result = temp_model.from_map(m['videoCaptionResult'])
|
|
@@ -10600,6 +10785,80 @@ class RunTagMiningAnalysisResponse(TeaModel):
|
|
|
10600
10785
|
return self
|
|
10601
10786
|
|
|
10602
10787
|
|
|
10788
|
+
class RunVideoAnalysisRequestAddDocumentParamDocument(TeaModel):
|
|
10789
|
+
def __init__(
|
|
10790
|
+
self,
|
|
10791
|
+
doc_id: str = None,
|
|
10792
|
+
title: str = None,
|
|
10793
|
+
):
|
|
10794
|
+
self.doc_id = doc_id
|
|
10795
|
+
self.title = title
|
|
10796
|
+
|
|
10797
|
+
def validate(self):
|
|
10798
|
+
pass
|
|
10799
|
+
|
|
10800
|
+
def to_map(self):
|
|
10801
|
+
_map = super().to_map()
|
|
10802
|
+
if _map is not None:
|
|
10803
|
+
return _map
|
|
10804
|
+
|
|
10805
|
+
result = dict()
|
|
10806
|
+
if self.doc_id is not None:
|
|
10807
|
+
result['docId'] = self.doc_id
|
|
10808
|
+
if self.title is not None:
|
|
10809
|
+
result['title'] = self.title
|
|
10810
|
+
return result
|
|
10811
|
+
|
|
10812
|
+
def from_map(self, m: dict = None):
|
|
10813
|
+
m = m or dict()
|
|
10814
|
+
if m.get('docId') is not None:
|
|
10815
|
+
self.doc_id = m.get('docId')
|
|
10816
|
+
if m.get('title') is not None:
|
|
10817
|
+
self.title = m.get('title')
|
|
10818
|
+
return self
|
|
10819
|
+
|
|
10820
|
+
|
|
10821
|
+
class RunVideoAnalysisRequestAddDocumentParam(TeaModel):
|
|
10822
|
+
def __init__(
|
|
10823
|
+
self,
|
|
10824
|
+
dataset_id: int = None,
|
|
10825
|
+
dataset_name: str = None,
|
|
10826
|
+
document: RunVideoAnalysisRequestAddDocumentParamDocument = None,
|
|
10827
|
+
):
|
|
10828
|
+
self.dataset_id = dataset_id
|
|
10829
|
+
self.dataset_name = dataset_name
|
|
10830
|
+
self.document = document
|
|
10831
|
+
|
|
10832
|
+
def validate(self):
|
|
10833
|
+
if self.document:
|
|
10834
|
+
self.document.validate()
|
|
10835
|
+
|
|
10836
|
+
def to_map(self):
|
|
10837
|
+
_map = super().to_map()
|
|
10838
|
+
if _map is not None:
|
|
10839
|
+
return _map
|
|
10840
|
+
|
|
10841
|
+
result = dict()
|
|
10842
|
+
if self.dataset_id is not None:
|
|
10843
|
+
result['datasetId'] = self.dataset_id
|
|
10844
|
+
if self.dataset_name is not None:
|
|
10845
|
+
result['datasetName'] = self.dataset_name
|
|
10846
|
+
if self.document is not None:
|
|
10847
|
+
result['document'] = self.document.to_map()
|
|
10848
|
+
return result
|
|
10849
|
+
|
|
10850
|
+
def from_map(self, m: dict = None):
|
|
10851
|
+
m = m or dict()
|
|
10852
|
+
if m.get('datasetId') is not None:
|
|
10853
|
+
self.dataset_id = m.get('datasetId')
|
|
10854
|
+
if m.get('datasetName') is not None:
|
|
10855
|
+
self.dataset_name = m.get('datasetName')
|
|
10856
|
+
if m.get('document') is not None:
|
|
10857
|
+
temp_model = RunVideoAnalysisRequestAddDocumentParamDocument()
|
|
10858
|
+
self.document = temp_model.from_map(m['document'])
|
|
10859
|
+
return self
|
|
10860
|
+
|
|
10861
|
+
|
|
10603
10862
|
class RunVideoAnalysisRequestFrameSampleMethod(TeaModel):
|
|
10604
10863
|
def __init__(
|
|
10605
10864
|
self,
|
|
@@ -10859,6 +11118,7 @@ class RunVideoAnalysisRequestVideoRoles(TeaModel):
|
|
|
10859
11118
|
class RunVideoAnalysisRequest(TeaModel):
|
|
10860
11119
|
def __init__(
|
|
10861
11120
|
self,
|
|
11121
|
+
add_document_param: RunVideoAnalysisRequestAddDocumentParam = None,
|
|
10862
11122
|
auto_role_recognition_video_url: str = None,
|
|
10863
11123
|
exclude_generate_options: List[str] = None,
|
|
10864
11124
|
face_identity_similarity_min_score: float = None,
|
|
@@ -10882,6 +11142,7 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
10882
11142
|
video_shot_face_identity_count: int = None,
|
|
10883
11143
|
video_url: str = None,
|
|
10884
11144
|
):
|
|
11145
|
+
self.add_document_param = add_document_param
|
|
10885
11146
|
self.auto_role_recognition_video_url = auto_role_recognition_video_url
|
|
10886
11147
|
self.exclude_generate_options = exclude_generate_options
|
|
10887
11148
|
self.face_identity_similarity_min_score = face_identity_similarity_min_score
|
|
@@ -10906,6 +11167,8 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
10906
11167
|
self.video_url = video_url
|
|
10907
11168
|
|
|
10908
11169
|
def validate(self):
|
|
11170
|
+
if self.add_document_param:
|
|
11171
|
+
self.add_document_param.validate()
|
|
10909
11172
|
if self.frame_sample_method:
|
|
10910
11173
|
self.frame_sample_method.validate()
|
|
10911
11174
|
if self.text_process_tasks:
|
|
@@ -10925,6 +11188,8 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
10925
11188
|
return _map
|
|
10926
11189
|
|
|
10927
11190
|
result = dict()
|
|
11191
|
+
if self.add_document_param is not None:
|
|
11192
|
+
result['addDocumentParam'] = self.add_document_param.to_map()
|
|
10928
11193
|
if self.auto_role_recognition_video_url is not None:
|
|
10929
11194
|
result['autoRoleRecognitionVideoUrl'] = self.auto_role_recognition_video_url
|
|
10930
11195
|
if self.exclude_generate_options is not None:
|
|
@@ -10977,6 +11242,9 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
10977
11242
|
|
|
10978
11243
|
def from_map(self, m: dict = None):
|
|
10979
11244
|
m = m or dict()
|
|
11245
|
+
if m.get('addDocumentParam') is not None:
|
|
11246
|
+
temp_model = RunVideoAnalysisRequestAddDocumentParam()
|
|
11247
|
+
self.add_document_param = temp_model.from_map(m['addDocumentParam'])
|
|
10980
11248
|
if m.get('autoRoleRecognitionVideoUrl') is not None:
|
|
10981
11249
|
self.auto_role_recognition_video_url = m.get('autoRoleRecognitionVideoUrl')
|
|
10982
11250
|
if m.get('excludeGenerateOptions') is not None:
|
|
@@ -11035,6 +11303,7 @@ class RunVideoAnalysisRequest(TeaModel):
|
|
|
11035
11303
|
class RunVideoAnalysisShrinkRequest(TeaModel):
|
|
11036
11304
|
def __init__(
|
|
11037
11305
|
self,
|
|
11306
|
+
add_document_param_shrink: str = None,
|
|
11038
11307
|
auto_role_recognition_video_url: str = None,
|
|
11039
11308
|
exclude_generate_options_shrink: str = None,
|
|
11040
11309
|
face_identity_similarity_min_score: float = None,
|
|
@@ -11058,6 +11327,7 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
|
|
|
11058
11327
|
video_shot_face_identity_count: int = None,
|
|
11059
11328
|
video_url: str = None,
|
|
11060
11329
|
):
|
|
11330
|
+
self.add_document_param_shrink = add_document_param_shrink
|
|
11061
11331
|
self.auto_role_recognition_video_url = auto_role_recognition_video_url
|
|
11062
11332
|
self.exclude_generate_options_shrink = exclude_generate_options_shrink
|
|
11063
11333
|
self.face_identity_similarity_min_score = face_identity_similarity_min_score
|
|
@@ -11090,6 +11360,8 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
|
|
|
11090
11360
|
return _map
|
|
11091
11361
|
|
|
11092
11362
|
result = dict()
|
|
11363
|
+
if self.add_document_param_shrink is not None:
|
|
11364
|
+
result['addDocumentParam'] = self.add_document_param_shrink
|
|
11093
11365
|
if self.auto_role_recognition_video_url is not None:
|
|
11094
11366
|
result['autoRoleRecognitionVideoUrl'] = self.auto_role_recognition_video_url
|
|
11095
11367
|
if self.exclude_generate_options_shrink is not None:
|
|
@@ -11138,6 +11410,8 @@ class RunVideoAnalysisShrinkRequest(TeaModel):
|
|
|
11138
11410
|
|
|
11139
11411
|
def from_map(self, m: dict = None):
|
|
11140
11412
|
m = m or dict()
|
|
11413
|
+
if m.get('addDocumentParam') is not None:
|
|
11414
|
+
self.add_document_param_shrink = m.get('addDocumentParam')
|
|
11141
11415
|
if m.get('autoRoleRecognitionVideoUrl') is not None:
|
|
11142
11416
|
self.auto_role_recognition_video_url = m.get('autoRoleRecognitionVideoUrl')
|
|
11143
11417
|
if m.get('excludeGenerateOptions') is not None:
|
|
@@ -11248,13 +11522,66 @@ class RunVideoAnalysisResponseBodyHeader(TeaModel):
|
|
|
11248
11522
|
return self
|
|
11249
11523
|
|
|
11250
11524
|
|
|
11525
|
+
class RunVideoAnalysisResponseBodyPayloadOutputAddDatasetDocumentsResult(TeaModel):
|
|
11526
|
+
def __init__(
|
|
11527
|
+
self,
|
|
11528
|
+
doc_id: str = None,
|
|
11529
|
+
doc_uuid: str = None,
|
|
11530
|
+
error_message: str = None,
|
|
11531
|
+
status: int = None,
|
|
11532
|
+
title: str = None,
|
|
11533
|
+
):
|
|
11534
|
+
self.doc_id = doc_id
|
|
11535
|
+
self.doc_uuid = doc_uuid
|
|
11536
|
+
self.error_message = error_message
|
|
11537
|
+
self.status = status
|
|
11538
|
+
self.title = title
|
|
11539
|
+
|
|
11540
|
+
def validate(self):
|
|
11541
|
+
pass
|
|
11542
|
+
|
|
11543
|
+
def to_map(self):
|
|
11544
|
+
_map = super().to_map()
|
|
11545
|
+
if _map is not None:
|
|
11546
|
+
return _map
|
|
11547
|
+
|
|
11548
|
+
result = dict()
|
|
11549
|
+
if self.doc_id is not None:
|
|
11550
|
+
result['docId'] = self.doc_id
|
|
11551
|
+
if self.doc_uuid is not None:
|
|
11552
|
+
result['docUuid'] = self.doc_uuid
|
|
11553
|
+
if self.error_message is not None:
|
|
11554
|
+
result['errorMessage'] = self.error_message
|
|
11555
|
+
if self.status is not None:
|
|
11556
|
+
result['status'] = self.status
|
|
11557
|
+
if self.title is not None:
|
|
11558
|
+
result['title'] = self.title
|
|
11559
|
+
return result
|
|
11560
|
+
|
|
11561
|
+
def from_map(self, m: dict = None):
|
|
11562
|
+
m = m or dict()
|
|
11563
|
+
if m.get('docId') is not None:
|
|
11564
|
+
self.doc_id = m.get('docId')
|
|
11565
|
+
if m.get('docUuid') is not None:
|
|
11566
|
+
self.doc_uuid = m.get('docUuid')
|
|
11567
|
+
if m.get('errorMessage') is not None:
|
|
11568
|
+
self.error_message = m.get('errorMessage')
|
|
11569
|
+
if m.get('status') is not None:
|
|
11570
|
+
self.status = m.get('status')
|
|
11571
|
+
if m.get('title') is not None:
|
|
11572
|
+
self.title = m.get('title')
|
|
11573
|
+
return self
|
|
11574
|
+
|
|
11575
|
+
|
|
11251
11576
|
class RunVideoAnalysisResponseBodyPayloadOutputVideoAnalysisResultUsage(TeaModel):
|
|
11252
11577
|
def __init__(
|
|
11253
11578
|
self,
|
|
11579
|
+
image_tokens: int = None,
|
|
11254
11580
|
input_tokens: int = None,
|
|
11255
11581
|
output_tokens: int = None,
|
|
11256
11582
|
total_tokens: int = None,
|
|
11257
11583
|
):
|
|
11584
|
+
self.image_tokens = image_tokens
|
|
11258
11585
|
self.input_tokens = input_tokens
|
|
11259
11586
|
self.output_tokens = output_tokens
|
|
11260
11587
|
self.total_tokens = total_tokens
|
|
@@ -11268,6 +11595,8 @@ class RunVideoAnalysisResponseBodyPayloadOutputVideoAnalysisResultUsage(TeaModel
|
|
|
11268
11595
|
return _map
|
|
11269
11596
|
|
|
11270
11597
|
result = dict()
|
|
11598
|
+
if self.image_tokens is not None:
|
|
11599
|
+
result['imageTokens'] = self.image_tokens
|
|
11271
11600
|
if self.input_tokens is not None:
|
|
11272
11601
|
result['inputTokens'] = self.input_tokens
|
|
11273
11602
|
if self.output_tokens is not None:
|
|
@@ -11278,6 +11607,8 @@ class RunVideoAnalysisResponseBodyPayloadOutputVideoAnalysisResultUsage(TeaModel
|
|
|
11278
11607
|
|
|
11279
11608
|
def from_map(self, m: dict = None):
|
|
11280
11609
|
m = m or dict()
|
|
11610
|
+
if m.get('imageTokens') is not None:
|
|
11611
|
+
self.image_tokens = m.get('imageTokens')
|
|
11281
11612
|
if m.get('inputTokens') is not None:
|
|
11282
11613
|
self.input_tokens = m.get('inputTokens')
|
|
11283
11614
|
if m.get('outputTokens') is not None:
|
|
@@ -11388,6 +11719,116 @@ class RunVideoAnalysisResponseBodyPayloadOutputVideoAnalysisResult(TeaModel):
|
|
|
11388
11719
|
return self
|
|
11389
11720
|
|
|
11390
11721
|
|
|
11722
|
+
class RunVideoAnalysisResponseBodyPayloadOutputVideoCalculatorResultItems(TeaModel):
|
|
11723
|
+
def __init__(
|
|
11724
|
+
self,
|
|
11725
|
+
input_expense: float = None,
|
|
11726
|
+
input_token: int = None,
|
|
11727
|
+
name: str = None,
|
|
11728
|
+
output_expense: float = None,
|
|
11729
|
+
output_token: int = None,
|
|
11730
|
+
time: int = None,
|
|
11731
|
+
time_expense: float = None,
|
|
11732
|
+
total_expense: float = None,
|
|
11733
|
+
type: str = None,
|
|
11734
|
+
):
|
|
11735
|
+
self.input_expense = input_expense
|
|
11736
|
+
self.input_token = input_token
|
|
11737
|
+
self.name = name
|
|
11738
|
+
self.output_expense = output_expense
|
|
11739
|
+
self.output_token = output_token
|
|
11740
|
+
self.time = time
|
|
11741
|
+
self.time_expense = time_expense
|
|
11742
|
+
self.total_expense = total_expense
|
|
11743
|
+
self.type = type
|
|
11744
|
+
|
|
11745
|
+
def validate(self):
|
|
11746
|
+
pass
|
|
11747
|
+
|
|
11748
|
+
def to_map(self):
|
|
11749
|
+
_map = super().to_map()
|
|
11750
|
+
if _map is not None:
|
|
11751
|
+
return _map
|
|
11752
|
+
|
|
11753
|
+
result = dict()
|
|
11754
|
+
if self.input_expense is not None:
|
|
11755
|
+
result['inputExpense'] = self.input_expense
|
|
11756
|
+
if self.input_token is not None:
|
|
11757
|
+
result['inputToken'] = self.input_token
|
|
11758
|
+
if self.name is not None:
|
|
11759
|
+
result['name'] = self.name
|
|
11760
|
+
if self.output_expense is not None:
|
|
11761
|
+
result['outputExpense'] = self.output_expense
|
|
11762
|
+
if self.output_token is not None:
|
|
11763
|
+
result['outputToken'] = self.output_token
|
|
11764
|
+
if self.time is not None:
|
|
11765
|
+
result['time'] = self.time
|
|
11766
|
+
if self.time_expense is not None:
|
|
11767
|
+
result['timeExpense'] = self.time_expense
|
|
11768
|
+
if self.total_expense is not None:
|
|
11769
|
+
result['totalExpense'] = self.total_expense
|
|
11770
|
+
if self.type is not None:
|
|
11771
|
+
result['type'] = self.type
|
|
11772
|
+
return result
|
|
11773
|
+
|
|
11774
|
+
def from_map(self, m: dict = None):
|
|
11775
|
+
m = m or dict()
|
|
11776
|
+
if m.get('inputExpense') is not None:
|
|
11777
|
+
self.input_expense = m.get('inputExpense')
|
|
11778
|
+
if m.get('inputToken') is not None:
|
|
11779
|
+
self.input_token = m.get('inputToken')
|
|
11780
|
+
if m.get('name') is not None:
|
|
11781
|
+
self.name = m.get('name')
|
|
11782
|
+
if m.get('outputExpense') is not None:
|
|
11783
|
+
self.output_expense = m.get('outputExpense')
|
|
11784
|
+
if m.get('outputToken') is not None:
|
|
11785
|
+
self.output_token = m.get('outputToken')
|
|
11786
|
+
if m.get('time') is not None:
|
|
11787
|
+
self.time = m.get('time')
|
|
11788
|
+
if m.get('timeExpense') is not None:
|
|
11789
|
+
self.time_expense = m.get('timeExpense')
|
|
11790
|
+
if m.get('totalExpense') is not None:
|
|
11791
|
+
self.total_expense = m.get('totalExpense')
|
|
11792
|
+
if m.get('type') is not None:
|
|
11793
|
+
self.type = m.get('type')
|
|
11794
|
+
return self
|
|
11795
|
+
|
|
11796
|
+
|
|
11797
|
+
class RunVideoAnalysisResponseBodyPayloadOutputVideoCalculatorResult(TeaModel):
|
|
11798
|
+
def __init__(
|
|
11799
|
+
self,
|
|
11800
|
+
items: List[RunVideoAnalysisResponseBodyPayloadOutputVideoCalculatorResultItems] = None,
|
|
11801
|
+
):
|
|
11802
|
+
self.items = items
|
|
11803
|
+
|
|
11804
|
+
def validate(self):
|
|
11805
|
+
if self.items:
|
|
11806
|
+
for k in self.items:
|
|
11807
|
+
if k:
|
|
11808
|
+
k.validate()
|
|
11809
|
+
|
|
11810
|
+
def to_map(self):
|
|
11811
|
+
_map = super().to_map()
|
|
11812
|
+
if _map is not None:
|
|
11813
|
+
return _map
|
|
11814
|
+
|
|
11815
|
+
result = dict()
|
|
11816
|
+
result['items'] = []
|
|
11817
|
+
if self.items is not None:
|
|
11818
|
+
for k in self.items:
|
|
11819
|
+
result['items'].append(k.to_map() if k else None)
|
|
11820
|
+
return result
|
|
11821
|
+
|
|
11822
|
+
def from_map(self, m: dict = None):
|
|
11823
|
+
m = m or dict()
|
|
11824
|
+
self.items = []
|
|
11825
|
+
if m.get('items') is not None:
|
|
11826
|
+
for k in m.get('items'):
|
|
11827
|
+
temp_model = RunVideoAnalysisResponseBodyPayloadOutputVideoCalculatorResultItems()
|
|
11828
|
+
self.items.append(temp_model.from_map(k))
|
|
11829
|
+
return self
|
|
11830
|
+
|
|
11831
|
+
|
|
11391
11832
|
class RunVideoAnalysisResponseBodyPayloadOutputVideoCaptionResultVideoCaptions(TeaModel):
|
|
11392
11833
|
def __init__(
|
|
11393
11834
|
self,
|
|
@@ -12259,8 +12700,10 @@ class RunVideoAnalysisResponseBodyPayloadOutputVideoTitleGenerateResult(TeaModel
|
|
|
12259
12700
|
class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
|
|
12260
12701
|
def __init__(
|
|
12261
12702
|
self,
|
|
12703
|
+
add_dataset_documents_result: RunVideoAnalysisResponseBodyPayloadOutputAddDatasetDocumentsResult = None,
|
|
12262
12704
|
result_json_file_url: str = None,
|
|
12263
12705
|
video_analysis_result: RunVideoAnalysisResponseBodyPayloadOutputVideoAnalysisResult = None,
|
|
12706
|
+
video_calculator_result: RunVideoAnalysisResponseBodyPayloadOutputVideoCalculatorResult = None,
|
|
12264
12707
|
video_caption_result: RunVideoAnalysisResponseBodyPayloadOutputVideoCaptionResult = None,
|
|
12265
12708
|
video_generate_result: RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResult = None,
|
|
12266
12709
|
video_generate_results: List[RunVideoAnalysisResponseBodyPayloadOutputVideoGenerateResults] = None,
|
|
@@ -12269,8 +12712,10 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
|
|
|
12269
12712
|
video_shot_snapshot_result: RunVideoAnalysisResponseBodyPayloadOutputVideoShotSnapshotResult = None,
|
|
12270
12713
|
video_title_generate_result: RunVideoAnalysisResponseBodyPayloadOutputVideoTitleGenerateResult = None,
|
|
12271
12714
|
):
|
|
12715
|
+
self.add_dataset_documents_result = add_dataset_documents_result
|
|
12272
12716
|
self.result_json_file_url = result_json_file_url
|
|
12273
12717
|
self.video_analysis_result = video_analysis_result
|
|
12718
|
+
self.video_calculator_result = video_calculator_result
|
|
12274
12719
|
self.video_caption_result = video_caption_result
|
|
12275
12720
|
self.video_generate_result = video_generate_result
|
|
12276
12721
|
self.video_generate_results = video_generate_results
|
|
@@ -12280,8 +12725,12 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
|
|
|
12280
12725
|
self.video_title_generate_result = video_title_generate_result
|
|
12281
12726
|
|
|
12282
12727
|
def validate(self):
|
|
12728
|
+
if self.add_dataset_documents_result:
|
|
12729
|
+
self.add_dataset_documents_result.validate()
|
|
12283
12730
|
if self.video_analysis_result:
|
|
12284
12731
|
self.video_analysis_result.validate()
|
|
12732
|
+
if self.video_calculator_result:
|
|
12733
|
+
self.video_calculator_result.validate()
|
|
12285
12734
|
if self.video_caption_result:
|
|
12286
12735
|
self.video_caption_result.validate()
|
|
12287
12736
|
if self.video_generate_result:
|
|
@@ -12305,10 +12754,14 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
|
|
|
12305
12754
|
return _map
|
|
12306
12755
|
|
|
12307
12756
|
result = dict()
|
|
12757
|
+
if self.add_dataset_documents_result is not None:
|
|
12758
|
+
result['addDatasetDocumentsResult'] = self.add_dataset_documents_result.to_map()
|
|
12308
12759
|
if self.result_json_file_url is not None:
|
|
12309
12760
|
result['resultJsonFileUrl'] = self.result_json_file_url
|
|
12310
12761
|
if self.video_analysis_result is not None:
|
|
12311
12762
|
result['videoAnalysisResult'] = self.video_analysis_result.to_map()
|
|
12763
|
+
if self.video_calculator_result is not None:
|
|
12764
|
+
result['videoCalculatorResult'] = self.video_calculator_result.to_map()
|
|
12312
12765
|
if self.video_caption_result is not None:
|
|
12313
12766
|
result['videoCaptionResult'] = self.video_caption_result.to_map()
|
|
12314
12767
|
if self.video_generate_result is not None:
|
|
@@ -12329,11 +12782,17 @@ class RunVideoAnalysisResponseBodyPayloadOutput(TeaModel):
|
|
|
12329
12782
|
|
|
12330
12783
|
def from_map(self, m: dict = None):
|
|
12331
12784
|
m = m or dict()
|
|
12785
|
+
if m.get('addDatasetDocumentsResult') is not None:
|
|
12786
|
+
temp_model = RunVideoAnalysisResponseBodyPayloadOutputAddDatasetDocumentsResult()
|
|
12787
|
+
self.add_dataset_documents_result = temp_model.from_map(m['addDatasetDocumentsResult'])
|
|
12332
12788
|
if m.get('resultJsonFileUrl') is not None:
|
|
12333
12789
|
self.result_json_file_url = m.get('resultJsonFileUrl')
|
|
12334
12790
|
if m.get('videoAnalysisResult') is not None:
|
|
12335
12791
|
temp_model = RunVideoAnalysisResponseBodyPayloadOutputVideoAnalysisResult()
|
|
12336
12792
|
self.video_analysis_result = temp_model.from_map(m['videoAnalysisResult'])
|
|
12793
|
+
if m.get('videoCalculatorResult') is not None:
|
|
12794
|
+
temp_model = RunVideoAnalysisResponseBodyPayloadOutputVideoCalculatorResult()
|
|
12795
|
+
self.video_calculator_result = temp_model.from_map(m['videoCalculatorResult'])
|
|
12337
12796
|
if m.get('videoCaptionResult') is not None:
|
|
12338
12797
|
temp_model = RunVideoAnalysisResponseBodyPayloadOutputVideoCaptionResult()
|
|
12339
12798
|
self.video_caption_result = temp_model.from_map(m['videoCaptionResult'])
|
|
@@ -14161,6 +14620,80 @@ class SubmitTagMiningAnalysisTaskResponse(TeaModel):
|
|
|
14161
14620
|
return self
|
|
14162
14621
|
|
|
14163
14622
|
|
|
14623
|
+
class SubmitVideoAnalysisTaskRequestAddDocumentParamDocument(TeaModel):
|
|
14624
|
+
def __init__(
|
|
14625
|
+
self,
|
|
14626
|
+
doc_id: str = None,
|
|
14627
|
+
title: str = None,
|
|
14628
|
+
):
|
|
14629
|
+
self.doc_id = doc_id
|
|
14630
|
+
self.title = title
|
|
14631
|
+
|
|
14632
|
+
def validate(self):
|
|
14633
|
+
pass
|
|
14634
|
+
|
|
14635
|
+
def to_map(self):
|
|
14636
|
+
_map = super().to_map()
|
|
14637
|
+
if _map is not None:
|
|
14638
|
+
return _map
|
|
14639
|
+
|
|
14640
|
+
result = dict()
|
|
14641
|
+
if self.doc_id is not None:
|
|
14642
|
+
result['docId'] = self.doc_id
|
|
14643
|
+
if self.title is not None:
|
|
14644
|
+
result['title'] = self.title
|
|
14645
|
+
return result
|
|
14646
|
+
|
|
14647
|
+
def from_map(self, m: dict = None):
|
|
14648
|
+
m = m or dict()
|
|
14649
|
+
if m.get('docId') is not None:
|
|
14650
|
+
self.doc_id = m.get('docId')
|
|
14651
|
+
if m.get('title') is not None:
|
|
14652
|
+
self.title = m.get('title')
|
|
14653
|
+
return self
|
|
14654
|
+
|
|
14655
|
+
|
|
14656
|
+
class SubmitVideoAnalysisTaskRequestAddDocumentParam(TeaModel):
|
|
14657
|
+
def __init__(
|
|
14658
|
+
self,
|
|
14659
|
+
dataset_id: int = None,
|
|
14660
|
+
dataset_name: str = None,
|
|
14661
|
+
document: SubmitVideoAnalysisTaskRequestAddDocumentParamDocument = None,
|
|
14662
|
+
):
|
|
14663
|
+
self.dataset_id = dataset_id
|
|
14664
|
+
self.dataset_name = dataset_name
|
|
14665
|
+
self.document = document
|
|
14666
|
+
|
|
14667
|
+
def validate(self):
|
|
14668
|
+
if self.document:
|
|
14669
|
+
self.document.validate()
|
|
14670
|
+
|
|
14671
|
+
def to_map(self):
|
|
14672
|
+
_map = super().to_map()
|
|
14673
|
+
if _map is not None:
|
|
14674
|
+
return _map
|
|
14675
|
+
|
|
14676
|
+
result = dict()
|
|
14677
|
+
if self.dataset_id is not None:
|
|
14678
|
+
result['datasetId'] = self.dataset_id
|
|
14679
|
+
if self.dataset_name is not None:
|
|
14680
|
+
result['datasetName'] = self.dataset_name
|
|
14681
|
+
if self.document is not None:
|
|
14682
|
+
result['document'] = self.document.to_map()
|
|
14683
|
+
return result
|
|
14684
|
+
|
|
14685
|
+
def from_map(self, m: dict = None):
|
|
14686
|
+
m = m or dict()
|
|
14687
|
+
if m.get('datasetId') is not None:
|
|
14688
|
+
self.dataset_id = m.get('datasetId')
|
|
14689
|
+
if m.get('datasetName') is not None:
|
|
14690
|
+
self.dataset_name = m.get('datasetName')
|
|
14691
|
+
if m.get('document') is not None:
|
|
14692
|
+
temp_model = SubmitVideoAnalysisTaskRequestAddDocumentParamDocument()
|
|
14693
|
+
self.document = temp_model.from_map(m['document'])
|
|
14694
|
+
return self
|
|
14695
|
+
|
|
14696
|
+
|
|
14164
14697
|
class SubmitVideoAnalysisTaskRequestFrameSampleMethod(TeaModel):
|
|
14165
14698
|
def __init__(
|
|
14166
14699
|
self,
|
|
@@ -14420,6 +14953,7 @@ class SubmitVideoAnalysisTaskRequestVideoRoles(TeaModel):
|
|
|
14420
14953
|
class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
14421
14954
|
def __init__(
|
|
14422
14955
|
self,
|
|
14956
|
+
add_document_param: SubmitVideoAnalysisTaskRequestAddDocumentParam = None,
|
|
14423
14957
|
auto_role_recognition_video_url: str = None,
|
|
14424
14958
|
deduplication_id: str = None,
|
|
14425
14959
|
exclude_generate_options: List[str] = None,
|
|
@@ -14442,6 +14976,7 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
14442
14976
|
video_shot_face_identity_count: int = None,
|
|
14443
14977
|
video_url: str = None,
|
|
14444
14978
|
):
|
|
14979
|
+
self.add_document_param = add_document_param
|
|
14445
14980
|
self.auto_role_recognition_video_url = auto_role_recognition_video_url
|
|
14446
14981
|
self.deduplication_id = deduplication_id
|
|
14447
14982
|
self.exclude_generate_options = exclude_generate_options
|
|
@@ -14466,6 +15001,8 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
14466
15001
|
self.video_url = video_url
|
|
14467
15002
|
|
|
14468
15003
|
def validate(self):
|
|
15004
|
+
if self.add_document_param:
|
|
15005
|
+
self.add_document_param.validate()
|
|
14469
15006
|
if self.frame_sample_method:
|
|
14470
15007
|
self.frame_sample_method.validate()
|
|
14471
15008
|
if self.text_process_tasks:
|
|
@@ -14485,6 +15022,8 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
14485
15022
|
return _map
|
|
14486
15023
|
|
|
14487
15024
|
result = dict()
|
|
15025
|
+
if self.add_document_param is not None:
|
|
15026
|
+
result['addDocumentParam'] = self.add_document_param.to_map()
|
|
14488
15027
|
if self.auto_role_recognition_video_url is not None:
|
|
14489
15028
|
result['autoRoleRecognitionVideoUrl'] = self.auto_role_recognition_video_url
|
|
14490
15029
|
if self.deduplication_id is not None:
|
|
@@ -14535,6 +15074,9 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
14535
15074
|
|
|
14536
15075
|
def from_map(self, m: dict = None):
|
|
14537
15076
|
m = m or dict()
|
|
15077
|
+
if m.get('addDocumentParam') is not None:
|
|
15078
|
+
temp_model = SubmitVideoAnalysisTaskRequestAddDocumentParam()
|
|
15079
|
+
self.add_document_param = temp_model.from_map(m['addDocumentParam'])
|
|
14538
15080
|
if m.get('autoRoleRecognitionVideoUrl') is not None:
|
|
14539
15081
|
self.auto_role_recognition_video_url = m.get('autoRoleRecognitionVideoUrl')
|
|
14540
15082
|
if m.get('deduplicationId') is not None:
|
|
@@ -14591,6 +15133,7 @@ class SubmitVideoAnalysisTaskRequest(TeaModel):
|
|
|
14591
15133
|
class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
|
|
14592
15134
|
def __init__(
|
|
14593
15135
|
self,
|
|
15136
|
+
add_document_param_shrink: str = None,
|
|
14594
15137
|
auto_role_recognition_video_url: str = None,
|
|
14595
15138
|
deduplication_id: str = None,
|
|
14596
15139
|
exclude_generate_options_shrink: str = None,
|
|
@@ -14613,6 +15156,7 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
|
|
|
14613
15156
|
video_shot_face_identity_count: int = None,
|
|
14614
15157
|
video_url: str = None,
|
|
14615
15158
|
):
|
|
15159
|
+
self.add_document_param_shrink = add_document_param_shrink
|
|
14616
15160
|
self.auto_role_recognition_video_url = auto_role_recognition_video_url
|
|
14617
15161
|
self.deduplication_id = deduplication_id
|
|
14618
15162
|
self.exclude_generate_options_shrink = exclude_generate_options_shrink
|
|
@@ -14645,6 +15189,8 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
|
|
|
14645
15189
|
return _map
|
|
14646
15190
|
|
|
14647
15191
|
result = dict()
|
|
15192
|
+
if self.add_document_param_shrink is not None:
|
|
15193
|
+
result['addDocumentParam'] = self.add_document_param_shrink
|
|
14648
15194
|
if self.auto_role_recognition_video_url is not None:
|
|
14649
15195
|
result['autoRoleRecognitionVideoUrl'] = self.auto_role_recognition_video_url
|
|
14650
15196
|
if self.deduplication_id is not None:
|
|
@@ -14691,6 +15237,8 @@ class SubmitVideoAnalysisTaskShrinkRequest(TeaModel):
|
|
|
14691
15237
|
|
|
14692
15238
|
def from_map(self, m: dict = None):
|
|
14693
15239
|
m = m or dict()
|
|
15240
|
+
if m.get('addDocumentParam') is not None:
|
|
15241
|
+
self.add_document_param_shrink = m.get('addDocumentParam')
|
|
14694
15242
|
if m.get('autoRoleRecognitionVideoUrl') is not None:
|
|
14695
15243
|
self.auto_role_recognition_video_url = m.get('autoRoleRecognitionVideoUrl')
|
|
14696
15244
|
if m.get('deduplicationId') is not None:
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.1
|
|
2
2
|
Name: alibabacloud-quanmiaolightapp20240801
|
|
3
|
-
Version: 2.13.
|
|
3
|
+
Version: 2.13.2
|
|
4
4
|
Summary: Alibaba Cloud QuanMiaoLightApp (20240801) SDK Library for Python
|
|
5
5
|
Home-page: https://github.com/aliyun/alibabacloud-python-sdk
|
|
6
6
|
Author: Alibaba Cloud SDK
|
alibabacloud_quanmiaolightapp20240801-2.13.0/alibabacloud_quanmiaolightapp20240801/__init__.py
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
__version__ = '2.13.0'
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|