tccli 3.0.1150.1__py2.py3-none-any.whl → 3.0.1151.1__py2.py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (31) hide show
  1. tccli/__init__.py +1 -1
  2. tccli/services/__init__.py +3 -0
  3. tccli/services/asr/v20190614/api.json +2 -2
  4. tccli/services/cbs/cbs_client.py +61 -8
  5. tccli/services/cbs/v20170312/api.json +64 -0
  6. tccli/services/cbs/v20170312/examples.json +8 -0
  7. tccli/services/ccc/v20200210/api.json +2 -2
  8. tccli/services/cdz/__init__.py +4 -0
  9. tccli/services/cdz/cdz_client.py +248 -0
  10. tccli/services/cdz/v20221123/api.json +305 -0
  11. tccli/services/cdz/v20221123/examples.json +21 -0
  12. tccli/services/clb/v20180317/api.json +9 -0
  13. tccli/services/cynosdb/v20190107/api.json +1 -1
  14. tccli/services/dasb/v20191018/api.json +5 -4
  15. tccli/services/ess/v20201111/api.json +4 -4
  16. tccli/services/essbasic/v20210526/api.json +2 -2
  17. tccli/services/essbasic/v20210526/examples.json +1 -1
  18. tccli/services/facefusion/v20220927/api.json +8 -8
  19. tccli/services/hunyuan/hunyuan_client.py +65 -12
  20. tccli/services/hunyuan/v20230901/api.json +89 -0
  21. tccli/services/hunyuan/v20230901/examples.json +8 -0
  22. tccli/services/mps/v20190612/api.json +37 -35
  23. tccli/services/mps/v20190612/examples.json +19 -7
  24. tccli/services/vpc/v20170312/api.json +1357 -57
  25. tccli/services/vpc/v20170312/examples.json +104 -0
  26. tccli/services/vpc/vpc_client.py +1974 -1285
  27. {tccli-3.0.1150.1.dist-info → tccli-3.0.1151.1.dist-info}/METADATA +2 -2
  28. {tccli-3.0.1150.1.dist-info → tccli-3.0.1151.1.dist-info}/RECORD +31 -27
  29. {tccli-3.0.1150.1.dist-info → tccli-3.0.1151.1.dist-info}/WHEEL +0 -0
  30. {tccli-3.0.1150.1.dist-info → tccli-3.0.1151.1.dist-info}/entry_points.txt +0 -0
  31. {tccli-3.0.1150.1.dist-info → tccli-3.0.1151.1.dist-info}/license_files/LICENSE +0 -0
@@ -17,7 +17,7 @@ from tencentcloud.hunyuan.v20230901 import models as models_v20230901
17
17
  from jmespath import search
18
18
  import time
19
19
 
20
- def doChatCompletions(args, parsed_globals):
20
+ def doSubmitHunyuanImageJob(args, parsed_globals):
21
21
  g_param = parse_global_arg(parsed_globals)
22
22
 
23
23
  if g_param[OptionsDefine.UseCVMRole.replace('-', '_')]:
@@ -46,11 +46,11 @@ def doChatCompletions(args, parsed_globals):
46
46
  client = mod.HunyuanClient(cred, g_param[OptionsDefine.Region], profile)
47
47
  client._sdkVersion += ("_CLI_" + __version__)
48
48
  models = MODELS_MAP[g_param[OptionsDefine.Version]]
49
- model = models.ChatCompletionsRequest()
49
+ model = models.SubmitHunyuanImageJobRequest()
50
50
  model.from_json_string(json.dumps(args))
51
51
  start_time = time.time()
52
52
  while True:
53
- rsp = client.ChatCompletions(model)
53
+ rsp = client.SubmitHunyuanImageJob(model)
54
54
  result = rsp.to_json_string()
55
55
  try:
56
56
  json_obj = json.loads(result)
@@ -121,7 +121,7 @@ def doQueryHunyuanImageJob(args, parsed_globals):
121
121
  FormatOutput.output("action", json_obj, g_param[OptionsDefine.Output], g_param[OptionsDefine.Filter])
122
122
 
123
123
 
124
- def doGetEmbedding(args, parsed_globals):
124
+ def doChatCompletions(args, parsed_globals):
125
125
  g_param = parse_global_arg(parsed_globals)
126
126
 
127
127
  if g_param[OptionsDefine.UseCVMRole.replace('-', '_')]:
@@ -150,11 +150,11 @@ def doGetEmbedding(args, parsed_globals):
150
150
  client = mod.HunyuanClient(cred, g_param[OptionsDefine.Region], profile)
151
151
  client._sdkVersion += ("_CLI_" + __version__)
152
152
  models = MODELS_MAP[g_param[OptionsDefine.Version]]
153
- model = models.GetEmbeddingRequest()
153
+ model = models.ChatCompletionsRequest()
154
154
  model.from_json_string(json.dumps(args))
155
155
  start_time = time.time()
156
156
  while True:
157
- rsp = client.GetEmbedding(model)
157
+ rsp = client.ChatCompletions(model)
158
158
  result = rsp.to_json_string()
159
159
  try:
160
160
  json_obj = json.loads(result)
@@ -225,7 +225,7 @@ def doGetTokenCount(args, parsed_globals):
225
225
  FormatOutput.output("action", json_obj, g_param[OptionsDefine.Output], g_param[OptionsDefine.Filter])
226
226
 
227
227
 
228
- def doSubmitHunyuanImageJob(args, parsed_globals):
228
+ def doTextToImageLite(args, parsed_globals):
229
229
  g_param = parse_global_arg(parsed_globals)
230
230
 
231
231
  if g_param[OptionsDefine.UseCVMRole.replace('-', '_')]:
@@ -254,11 +254,63 @@ def doSubmitHunyuanImageJob(args, parsed_globals):
254
254
  client = mod.HunyuanClient(cred, g_param[OptionsDefine.Region], profile)
255
255
  client._sdkVersion += ("_CLI_" + __version__)
256
256
  models = MODELS_MAP[g_param[OptionsDefine.Version]]
257
- model = models.SubmitHunyuanImageJobRequest()
257
+ model = models.TextToImageLiteRequest()
258
258
  model.from_json_string(json.dumps(args))
259
259
  start_time = time.time()
260
260
  while True:
261
- rsp = client.SubmitHunyuanImageJob(model)
261
+ rsp = client.TextToImageLite(model)
262
+ result = rsp.to_json_string()
263
+ try:
264
+ json_obj = json.loads(result)
265
+ except TypeError as e:
266
+ json_obj = json.loads(result.decode('utf-8')) # python3.3
267
+ if not g_param[OptionsDefine.Waiter] or search(g_param['OptionsDefine.WaiterInfo']['expr'], json_obj) == g_param['OptionsDefine.WaiterInfo']['to']:
268
+ break
269
+ cur_time = time.time()
270
+ if cur_time - start_time >= g_param['OptionsDefine.WaiterInfo']['timeout']:
271
+ raise ClientError('Request timeout, wait `%s` to `%s` timeout, last request is %s' %
272
+ (g_param['OptionsDefine.WaiterInfo']['expr'], g_param['OptionsDefine.WaiterInfo']['to'],
273
+ search(g_param['OptionsDefine.WaiterInfo']['expr'], json_obj)))
274
+ else:
275
+ print('Inquiry result is %s.' % search(g_param['OptionsDefine.WaiterInfo']['expr'], json_obj))
276
+ time.sleep(g_param['OptionsDefine.WaiterInfo']['interval'])
277
+ FormatOutput.output("action", json_obj, g_param[OptionsDefine.Output], g_param[OptionsDefine.Filter])
278
+
279
+
280
+ def doGetEmbedding(args, parsed_globals):
281
+ g_param = parse_global_arg(parsed_globals)
282
+
283
+ if g_param[OptionsDefine.UseCVMRole.replace('-', '_')]:
284
+ cred = credential.CVMRoleCredential()
285
+ elif g_param[OptionsDefine.RoleArn.replace('-', '_')] and g_param[OptionsDefine.RoleSessionName.replace('-', '_')]:
286
+ cred = credential.STSAssumeRoleCredential(
287
+ g_param[OptionsDefine.SecretId], g_param[OptionsDefine.SecretKey], g_param[OptionsDefine.RoleArn.replace('-', '_')],
288
+ g_param[OptionsDefine.RoleSessionName.replace('-', '_')], endpoint=g_param["sts_cred_endpoint"]
289
+ )
290
+ elif os.getenv(OptionsDefine.ENV_TKE_REGION) and os.getenv(OptionsDefine.ENV_TKE_PROVIDER_ID) and os.getenv(OptionsDefine.ENV_TKE_WEB_IDENTITY_TOKEN_FILE) and os.getenv(OptionsDefine.ENV_TKE_ROLE_ARN):
291
+ cred = credential.DefaultTkeOIDCRoleArnProvider().get_credentials()
292
+ else:
293
+ cred = credential.Credential(
294
+ g_param[OptionsDefine.SecretId], g_param[OptionsDefine.SecretKey], g_param[OptionsDefine.Token]
295
+ )
296
+ http_profile = HttpProfile(
297
+ reqTimeout=60 if g_param[OptionsDefine.Timeout] is None else int(g_param[OptionsDefine.Timeout]),
298
+ reqMethod="POST",
299
+ endpoint=g_param[OptionsDefine.Endpoint],
300
+ proxy=g_param[OptionsDefine.HttpsProxy.replace('-', '_')]
301
+ )
302
+ profile = ClientProfile(httpProfile=http_profile, signMethod="HmacSHA256")
303
+ if g_param[OptionsDefine.Language]:
304
+ profile.language = g_param[OptionsDefine.Language]
305
+ mod = CLIENT_MAP[g_param[OptionsDefine.Version]]
306
+ client = mod.HunyuanClient(cred, g_param[OptionsDefine.Region], profile)
307
+ client._sdkVersion += ("_CLI_" + __version__)
308
+ models = MODELS_MAP[g_param[OptionsDefine.Version]]
309
+ model = models.GetEmbeddingRequest()
310
+ model.from_json_string(json.dumps(args))
311
+ start_time = time.time()
312
+ while True:
313
+ rsp = client.GetEmbedding(model)
262
314
  result = rsp.to_json_string()
263
315
  try:
264
316
  json_obj = json.loads(result)
@@ -288,11 +340,12 @@ MODELS_MAP = {
288
340
  }
289
341
 
290
342
  ACTION_MAP = {
291
- "ChatCompletions": doChatCompletions,
343
+ "SubmitHunyuanImageJob": doSubmitHunyuanImageJob,
292
344
  "QueryHunyuanImageJob": doQueryHunyuanImageJob,
293
- "GetEmbedding": doGetEmbedding,
345
+ "ChatCompletions": doChatCompletions,
294
346
  "GetTokenCount": doGetTokenCount,
295
- "SubmitHunyuanImageJob": doSubmitHunyuanImageJob,
347
+ "TextToImageLite": doTextToImageLite,
348
+ "GetEmbedding": doGetEmbedding,
296
349
 
297
350
  }
298
351
 
@@ -34,6 +34,13 @@
34
34
  "name": "提交混元生图任务",
35
35
  "output": "SubmitHunyuanImageJobResponse",
36
36
  "status": "online"
37
+ },
38
+ "TextToImageLite": {
39
+ "document": "文生图轻量版接口根据输入的文本描述,智能生成与之相关的结果图。\n文生图轻量版默认提供3个并发任务数,代表最多能同时处理3个已提交的任务,上一个任务处理完毕后才能开始处理下一个任务。",
40
+ "input": "TextToImageLiteRequest",
41
+ "name": "文生图轻量版",
42
+ "output": "TextToImageLiteResponse",
43
+ "status": "online"
37
44
  }
38
45
  },
39
46
  "metadata": {
@@ -762,6 +769,88 @@
762
769
  ],
763
770
  "type": "object"
764
771
  },
772
+ "TextToImageLiteRequest": {
773
+ "document": "TextToImageLite请求参数结构体",
774
+ "members": [
775
+ {
776
+ "disabled": false,
777
+ "document": "文本描述。\n算法将根据输入的文本智能生成与之相关的图像。建议详细描述画面主体、细节、场景等,文本描述越丰富,生成效果越精美。\n不能为空,推荐使用中文。最多可传256个 utf-8 字符。",
778
+ "example": "雨中, 竹林, 小路",
779
+ "member": "string",
780
+ "name": "Prompt",
781
+ "required": true,
782
+ "type": "string"
783
+ },
784
+ {
785
+ "disabled": false,
786
+ "document": "反向文本描述。\n用于一定程度上从反面引导模型生成的走向,减少生成结果中出现描述内容的可能,但不能完全杜绝。\n推荐使用中文。最多可传256个 utf-8 字符。",
787
+ "example": "无",
788
+ "member": "string",
789
+ "name": "NegativePrompt",
790
+ "required": false,
791
+ "type": "string"
792
+ },
793
+ {
794
+ "disabled": false,
795
+ "document": "绘画风格。\n请在 [智能文生图风格列表](https://cloud.tencent.com/document/product/1668/86249) 中选择期望的风格,传入风格编号。\n推荐使用且只使用一种风格。不传默认使用201(日系动漫风格)。",
796
+ "example": "无",
797
+ "member": "string",
798
+ "name": "Style",
799
+ "required": false,
800
+ "type": "string"
801
+ },
802
+ {
803
+ "disabled": false,
804
+ "document": "生成图分辨率。\n支持生成以下分辨率的图片:768:768(1:1)、768:1024(3:4)、1024:768(4:3)、1024:1024(1:1)、720:1280(9:16)、1280:720(16:9)、768:1280(3:5)、1280:768(5:3)、1080:1920(9:16)、1920:1080(16:9),不传默认使用768:768。",
805
+ "example": "无",
806
+ "member": "string",
807
+ "name": "Resolution",
808
+ "required": false,
809
+ "type": "string"
810
+ },
811
+ {
812
+ "disabled": false,
813
+ "document": "为生成结果图添加标识的开关,默认为1。\n1:添加标识。\n0:不添加标识。\n其他数值:默认按0处理。\n建议您使用显著标识来提示结果图使用了 AI 绘画技术,是 AI 生成的图片。",
814
+ "example": "无",
815
+ "member": "int64",
816
+ "name": "LogoAdd",
817
+ "required": false,
818
+ "type": "int"
819
+ },
820
+ {
821
+ "disabled": false,
822
+ "document": "返回图像方式(base64 或 url) ,二选一,默认为 base64。url 有效期为1小时。",
823
+ "example": "base64",
824
+ "member": "string",
825
+ "name": "RspImgType",
826
+ "required": false,
827
+ "type": "string"
828
+ }
829
+ ],
830
+ "type": "object"
831
+ },
832
+ "TextToImageLiteResponse": {
833
+ "document": "TextToImageLite返回参数结构体",
834
+ "members": [
835
+ {
836
+ "disabled": false,
837
+ "document": "根据入参 RspImgType 填入不同,返回不同的内容。如果传入 base64 则返回生成图 Base64 编码。如果传入 url 则返回的生成图 URL , 有效期1小时,请及时保存。",
838
+ "example": "url",
839
+ "member": "string",
840
+ "name": "ResultImage",
841
+ "output_required": true,
842
+ "type": "string",
843
+ "value_allowed_null": false
844
+ },
845
+ {
846
+ "document": "唯一请求 ID,由服务端生成,每次请求都会返回(若请求因其他原因未能抵达服务端,则该次请求不会获得 RequestId)。定位问题时需要提供该次请求的 RequestId。",
847
+ "member": "string",
848
+ "name": "RequestId",
849
+ "type": "string"
850
+ }
851
+ ],
852
+ "type": "object"
853
+ },
765
854
  "Tool": {
766
855
  "document": "用户指定模型使用的工具",
767
856
  "members": [
@@ -87,6 +87,14 @@
87
87
  "output": "{\n \"Response\": {\n \"JobId\": \"test\",\n \"RequestId\": \"d4e1f224-fa21-40bc-9ee7-4bb13abece6e\"\n }\n}",
88
88
  "title": "调用成功"
89
89
  }
90
+ ],
91
+ "TextToImageLite": [
92
+ {
93
+ "document": "成功调用,并返回生图url",
94
+ "input": "POST / HTTP/1.1\nHost: hunyuan.tencentcloudapi.com\nContent-Type: application/json\nX-TC-Action: TextToImageLite\n<公共请求参数>\n\n{\n \"Prompt\": \"冰川,企鹅\",\n \"RspImgType\": \"url\"\n}",
95
+ "output": "{\n \"Response\": {\n \"RequestId\": \"53549a10-0633-4ebe-bf39-466dc51aa2bb\",\n \"ResultImage\": \"https://rspimg.jpg\"\n }\n}",
96
+ "title": "成功调用"
97
+ }
90
98
  ]
91
99
  },
92
100
  "version": "1.0"
@@ -127,7 +127,7 @@
127
127
  "status": "online"
128
128
  },
129
129
  "CreateTranscodeTemplate": {
130
- "document": "创建用户自定义转码模板,数量上限:1000",
130
+ "document": "创建用户自定义转码模板,数量上限:1000",
131
131
  "input": "CreateTranscodeTemplateRequest",
132
132
  "name": "创建转码模板",
133
133
  "output": "CreateTranscodeTemplateResponse",
@@ -8137,7 +8137,7 @@
8137
8137
  "example": "1008",
8138
8138
  "member": "uint64",
8139
8139
  "name": "Definition",
8140
- "required": true,
8140
+ "output_required": true,
8141
8141
  "type": "int",
8142
8142
  "value_allowed_null": false
8143
8143
  },
@@ -8228,7 +8228,7 @@
8228
8228
  "example": "30",
8229
8229
  "member": "int64",
8230
8230
  "name": "Definition",
8231
- "required": true,
8231
+ "output_required": true,
8232
8232
  "type": "int",
8233
8233
  "value_allowed_null": false
8234
8234
  },
@@ -8355,7 +8355,7 @@
8355
8355
  "example": "1008",
8356
8356
  "member": "uint64",
8357
8357
  "name": "Definition",
8358
- "required": true,
8358
+ "output_required": true,
8359
8359
  "type": "int",
8360
8360
  "value_allowed_null": false
8361
8361
  },
@@ -9042,6 +9042,7 @@
9042
9042
  "example": "",
9043
9043
  "member": "AiSamplePerson",
9044
9044
  "name": "Person",
9045
+ "output_required": true,
9045
9046
  "type": "object",
9046
9047
  "value_allowed_null": false
9047
9048
  },
@@ -9051,6 +9052,7 @@
9051
9052
  "example": "",
9052
9053
  "member": "AiSampleFailFaceInfo",
9053
9054
  "name": "FailFaceInfoSet",
9055
+ "output_required": true,
9054
9056
  "type": "list",
9055
9057
  "value_allowed_null": false
9056
9058
  },
@@ -9214,7 +9216,7 @@
9214
9216
  "example": "1008",
9215
9217
  "member": "uint64",
9216
9218
  "name": "Definition",
9217
- "required": true,
9219
+ "output_required": true,
9218
9220
  "type": "int",
9219
9221
  "value_allowed_null": false
9220
9222
  },
@@ -9978,6 +9980,7 @@
9978
9980
  "example": "157482",
9979
9981
  "member": "int64",
9980
9982
  "name": "WorkflowId",
9983
+ "output_required": true,
9981
9984
  "type": "int",
9982
9985
  "value_allowed_null": false
9983
9986
  },
@@ -10437,7 +10440,7 @@
10437
10440
  {
10438
10441
  "disabled": false,
10439
10442
  "document": "关键词,数组长度限制:100 个词。",
10440
- "example": "",
10443
+ "example": "“”",
10441
10444
  "member": "string",
10442
10445
  "name": "Keywords",
10443
10446
  "required": true,
@@ -13966,7 +13969,7 @@
13966
13969
  {
13967
13970
  "disabled": false,
13968
13971
  "document": "工作流 ID。",
13969
- "example": "",
13972
+ "example": "12345",
13970
13973
  "member": "int64",
13971
13974
  "name": "WorkflowId",
13972
13975
  "required": true,
@@ -14435,7 +14438,7 @@
14435
14438
  "example": "XXX",
14436
14439
  "member": "string",
14437
14440
  "name": "Result",
14438
- "required": true,
14441
+ "output_required": true,
14439
14442
  "type": "string",
14440
14443
  "value_allowed_null": false
14441
14444
  },
@@ -20488,6 +20491,7 @@
20488
20491
  "example": "",
20489
20492
  "member": "AiSamplePerson",
20490
20493
  "name": "Person",
20494
+ "output_required": true,
20491
20495
  "type": "object",
20492
20496
  "value_allowed_null": false
20493
20497
  },
@@ -20497,6 +20501,7 @@
20497
20501
  "example": "",
20498
20502
  "member": "AiSampleFailFaceInfo",
20499
20503
  "name": "FailFaceInfoSet",
20504
+ "output_required": true,
20500
20505
  "type": "list",
20501
20506
  "value_allowed_null": true
20502
20507
  },
@@ -21228,7 +21233,7 @@
21228
21233
  "example": "\"\"",
21229
21234
  "member": "string",
21230
21235
  "name": "ImageUrl",
21231
- "required": true,
21236
+ "output_required": true,
21232
21237
  "type": "string",
21233
21238
  "value_allowed_null": false
21234
21239
  },
@@ -23565,6 +23570,7 @@
23565
23570
  "example": "",
23566
23571
  "member": "string",
23567
23572
  "name": "TaskId",
23573
+ "output_required": true,
23568
23574
  "type": "string",
23569
23575
  "value_allowed_null": false
23570
23576
  },
@@ -23599,7 +23605,7 @@
23599
23605
  {
23600
23606
  "disabled": false,
23601
23607
  "document": "工作流 ID。",
23602
- "example": "",
23608
+ "example": "10",
23603
23609
  "member": "int64",
23604
23610
  "name": "WorkflowId",
23605
23611
  "required": true,
@@ -23608,7 +23614,7 @@
23608
23614
  {
23609
23615
  "disabled": false,
23610
23616
  "document": "工作流名称,最多128字符。同一个用户该名称唯一。",
23611
- "example": "",
23617
+ "example": "“”",
23612
23618
  "member": "string",
23613
23619
  "name": "WorkflowName",
23614
23620
  "required": true,
@@ -23617,7 +23623,7 @@
23617
23623
  {
23618
23624
  "disabled": false,
23619
23625
  "document": "工作流绑定的触发规则,当上传视频命中该规则到该对象时即触发工作流。",
23620
- "example": "",
23626
+ "example": "",
23621
23627
  "member": "WorkflowTrigger",
23622
23628
  "name": "Trigger",
23623
23629
  "required": true,
@@ -23626,7 +23632,7 @@
23626
23632
  {
23627
23633
  "disabled": false,
23628
23634
  "document": "视频处理的文件输出配置。不填则继承 Trigger 中的存储位置。",
23629
- "example": "",
23635
+ "example": "",
23630
23636
  "member": "TaskOutputStorage",
23631
23637
  "name": "OutputStorage",
23632
23638
  "required": false,
@@ -23634,8 +23640,8 @@
23634
23640
  },
23635
23641
  {
23636
23642
  "disabled": false,
23637
- "document": "视频处理生成的文件输出的目标目录,如`/movie/201907/`。如果不填,表示与触发文件所在的目录一致,即`{inputDir}`。",
23638
- "example": "",
23643
+ "document": "视频处理生成的文件输出的目标目录,必选以 / 开头和结尾,如`/movie/201907/`。\n如果不填,表示与触发文件所在的目录一致,即`{inputDir}`。",
23644
+ "example": "",
23639
23645
  "member": "string",
23640
23646
  "name": "OutputDir",
23641
23647
  "required": false,
@@ -23644,7 +23650,7 @@
23644
23650
  {
23645
23651
  "disabled": false,
23646
23652
  "document": "视频处理类型任务参数。",
23647
- "example": "",
23653
+ "example": "",
23648
23654
  "member": "MediaProcessTaskInput",
23649
23655
  "name": "MediaProcessTask",
23650
23656
  "required": false,
@@ -23653,7 +23659,7 @@
23653
23659
  {
23654
23660
  "disabled": false,
23655
23661
  "document": "视频内容审核类型任务参数。",
23656
- "example": "",
23662
+ "example": "",
23657
23663
  "member": "AiContentReviewTaskInput",
23658
23664
  "name": "AiContentReviewTask",
23659
23665
  "required": false,
@@ -23662,7 +23668,7 @@
23662
23668
  {
23663
23669
  "disabled": false,
23664
23670
  "document": "视频内容分析类型任务参数。",
23665
- "example": "",
23671
+ "example": "",
23666
23672
  "member": "AiAnalysisTaskInput",
23667
23673
  "name": "AiAnalysisTask",
23668
23674
  "required": false,
@@ -23671,7 +23677,7 @@
23671
23677
  {
23672
23678
  "disabled": false,
23673
23679
  "document": "视频内容识别类型任务参数。",
23674
- "example": "",
23680
+ "example": "",
23675
23681
  "member": "AiRecognitionTaskInput",
23676
23682
  "name": "AiRecognitionTask",
23677
23683
  "required": false,
@@ -23680,7 +23686,7 @@
23680
23686
  {
23681
23687
  "disabled": false,
23682
23688
  "document": "工作流的优先级,数值越大优先级越高,取值范围是 -10 到 10,不填代表 0。",
23683
- "example": "",
23689
+ "example": "",
23684
23690
  "member": "int64",
23685
23691
  "name": "TaskPriority",
23686
23692
  "required": false,
@@ -23689,7 +23695,7 @@
23689
23695
  {
23690
23696
  "disabled": false,
23691
23697
  "document": "任务的事件通知信息,不填代表不获取事件通知。",
23692
- "example": "",
23698
+ "example": "",
23693
23699
  "member": "TaskNotifyConfig",
23694
23700
  "name": "TaskNotifyConfig",
23695
23701
  "required": false,
@@ -26539,42 +26545,38 @@
26539
26545
  {
26540
26546
  "disabled": false,
26541
26547
  "document": "用户自定文本审核任务开关,可选值:\n<li>ON:开启自定义文本审核任务;</li>\n<li>OFF:关闭自定义文本审核任务。</li>",
26542
- "example": "",
26548
+ "example": "",
26543
26549
  "member": "string",
26544
26550
  "name": "Switch",
26545
26551
  "required": false,
26546
- "type": "string",
26547
- "value_allowed_null": false
26552
+ "type": "string"
26548
26553
  },
26549
26554
  {
26550
26555
  "disabled": false,
26551
26556
  "document": "用户自定义文本过滤标签,审核结果包含选择的标签则返回结果,如果过滤标签为空,则审核结果全部返回。如果要使用标签过滤功能,添加自定义文本关键词素材时需要添加对应标签。\n标签个数最多 10 个,每个标签长度最多 16 个字符。",
26552
- "example": "",
26557
+ "example": "",
26553
26558
  "member": "string",
26554
26559
  "name": "LabelSet",
26555
26560
  "required": false,
26556
- "type": "string",
26557
- "value_allowed_null": false
26561
+ "type": "string"
26558
26562
  },
26559
26563
  {
26560
26564
  "disabled": false,
26561
26565
  "document": "判定涉嫌违规的分数阈值,当智能审核达到该分数以上,认为涉嫌违规。取值范围:0~100。",
26562
- "example": "",
26566
+ "example": "",
26563
26567
  "member": "int64",
26564
26568
  "name": "BlockConfidence",
26565
26569
  "required": false,
26566
- "type": "int",
26567
- "value_allowed_null": false
26570
+ "type": "int"
26568
26571
  },
26569
26572
  {
26570
26573
  "disabled": false,
26571
26574
  "document": "判定需人工复核是否违规的分数阈值,当智能审核达到该分数以上,认为需人工复核。取值范围:0~100。",
26572
- "example": "",
26575
+ "example": "",
26573
26576
  "member": "int64",
26574
26577
  "name": "ReviewConfidence",
26575
26578
  "required": false,
26576
- "type": "int",
26577
- "value_allowed_null": false
26579
+ "type": "int"
26578
26580
  }
26579
26581
  ],
26580
26582
  "usage": "in"
@@ -26737,7 +26739,7 @@
26737
26739
  },
26738
26740
  {
26739
26741
  "disabled": false,
26740
- "document": "视频帧率,取值范围:[0, 120],单位:Hz。 当取值为 0,表示帧率和原始视频保持一致。 注意:自适应码率时取值范围是 [0, 60]",
26742
+ "document": "视频帧率,取值范围:[0, 120],单位:Hz。 当取值为 0,表示帧率和原始视频保持一致。\n注意:自适应码率时取值范围是 [0, 60]",
26741
26743
  "example": "0",
26742
26744
  "member": "int64",
26743
26745
  "name": "Fps",
@@ -27232,7 +27234,7 @@
27232
27234
  "example": "125xxx65-WithdrawsWatermark-bffb15f07530b57bc1aabb01fac74bca",
27233
27235
  "member": "string",
27234
27236
  "name": "TaskId",
27235
- "required": true,
27237
+ "output_required": true,
27236
27238
  "type": "string",
27237
27239
  "value_allowed_null": false
27238
27240
  },