alibabacloud-ice20201109 6.8.3__py3-none-any.whl → 6.8.4__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- alibabacloud_ice20201109/__init__.py +1 -1
- alibabacloud_ice20201109/client.py +4 -0
- alibabacloud_ice20201109/models.py +426 -0
- {alibabacloud_ice20201109-6.8.3.dist-info → alibabacloud_ice20201109-6.8.4.dist-info}/METADATA +1 -1
- alibabacloud_ice20201109-6.8.4.dist-info/RECORD +8 -0
- alibabacloud_ice20201109-6.8.3.dist-info/RECORD +0 -8
- {alibabacloud_ice20201109-6.8.3.dist-info → alibabacloud_ice20201109-6.8.4.dist-info}/LICENSE +0 -0
- {alibabacloud_ice20201109-6.8.3.dist-info → alibabacloud_ice20201109-6.8.4.dist-info}/WHEEL +0 -0
- {alibabacloud_ice20201109-6.8.3.dist-info → alibabacloud_ice20201109-6.8.4.dist-info}/top_level.txt +0 -0
|
@@ -1 +1 @@
|
|
|
1
|
-
__version__ = '6.8.
|
|
1
|
+
__version__ = '6.8.4'
|
|
@@ -12363,6 +12363,8 @@ class Client(OpenApiClient):
|
|
|
12363
12363
|
query = {}
|
|
12364
12364
|
if not UtilClient.is_unset(request.called_number):
|
|
12365
12365
|
query['CalledNumber'] = request.called_number
|
|
12366
|
+
if not UtilClient.is_unset(request.caller_number):
|
|
12367
|
+
query['CallerNumber'] = request.caller_number
|
|
12366
12368
|
if not UtilClient.is_unset(request.error_prompt):
|
|
12367
12369
|
query['ErrorPrompt'] = request.error_prompt
|
|
12368
12370
|
if not UtilClient.is_unset(request.instance_id):
|
|
@@ -12404,6 +12406,8 @@ class Client(OpenApiClient):
|
|
|
12404
12406
|
query = {}
|
|
12405
12407
|
if not UtilClient.is_unset(request.called_number):
|
|
12406
12408
|
query['CalledNumber'] = request.called_number
|
|
12409
|
+
if not UtilClient.is_unset(request.caller_number):
|
|
12410
|
+
query['CallerNumber'] = request.caller_number
|
|
12407
12411
|
if not UtilClient.is_unset(request.error_prompt):
|
|
12408
12412
|
query['ErrorPrompt'] = request.error_prompt
|
|
12409
12413
|
if not UtilClient.is_unset(request.instance_id):
|
|
@@ -94,6 +94,198 @@ class AIAgentConfigAsrConfig(TeaModel):
|
|
|
94
94
|
return self
|
|
95
95
|
|
|
96
96
|
|
|
97
|
+
class AIAgentConfigAutoSpeechConfigLlmPendingMessages(TeaModel):
|
|
98
|
+
def __init__(
|
|
99
|
+
self,
|
|
100
|
+
probability: float = None,
|
|
101
|
+
text: str = None,
|
|
102
|
+
):
|
|
103
|
+
self.probability = probability
|
|
104
|
+
self.text = text
|
|
105
|
+
|
|
106
|
+
def validate(self):
|
|
107
|
+
pass
|
|
108
|
+
|
|
109
|
+
def to_map(self):
|
|
110
|
+
_map = super().to_map()
|
|
111
|
+
if _map is not None:
|
|
112
|
+
return _map
|
|
113
|
+
|
|
114
|
+
result = dict()
|
|
115
|
+
if self.probability is not None:
|
|
116
|
+
result['Probability'] = self.probability
|
|
117
|
+
if self.text is not None:
|
|
118
|
+
result['Text'] = self.text
|
|
119
|
+
return result
|
|
120
|
+
|
|
121
|
+
def from_map(self, m: dict = None):
|
|
122
|
+
m = m or dict()
|
|
123
|
+
if m.get('Probability') is not None:
|
|
124
|
+
self.probability = m.get('Probability')
|
|
125
|
+
if m.get('Text') is not None:
|
|
126
|
+
self.text = m.get('Text')
|
|
127
|
+
return self
|
|
128
|
+
|
|
129
|
+
|
|
130
|
+
class AIAgentConfigAutoSpeechConfigLlmPending(TeaModel):
|
|
131
|
+
def __init__(
|
|
132
|
+
self,
|
|
133
|
+
messages: List[AIAgentConfigAutoSpeechConfigLlmPendingMessages] = None,
|
|
134
|
+
wait_time: int = None,
|
|
135
|
+
):
|
|
136
|
+
self.messages = messages
|
|
137
|
+
self.wait_time = wait_time
|
|
138
|
+
|
|
139
|
+
def validate(self):
|
|
140
|
+
if self.messages:
|
|
141
|
+
for k in self.messages:
|
|
142
|
+
if k:
|
|
143
|
+
k.validate()
|
|
144
|
+
|
|
145
|
+
def to_map(self):
|
|
146
|
+
_map = super().to_map()
|
|
147
|
+
if _map is not None:
|
|
148
|
+
return _map
|
|
149
|
+
|
|
150
|
+
result = dict()
|
|
151
|
+
result['Messages'] = []
|
|
152
|
+
if self.messages is not None:
|
|
153
|
+
for k in self.messages:
|
|
154
|
+
result['Messages'].append(k.to_map() if k else None)
|
|
155
|
+
if self.wait_time is not None:
|
|
156
|
+
result['WaitTime'] = self.wait_time
|
|
157
|
+
return result
|
|
158
|
+
|
|
159
|
+
def from_map(self, m: dict = None):
|
|
160
|
+
m = m or dict()
|
|
161
|
+
self.messages = []
|
|
162
|
+
if m.get('Messages') is not None:
|
|
163
|
+
for k in m.get('Messages'):
|
|
164
|
+
temp_model = AIAgentConfigAutoSpeechConfigLlmPendingMessages()
|
|
165
|
+
self.messages.append(temp_model.from_map(k))
|
|
166
|
+
if m.get('WaitTime') is not None:
|
|
167
|
+
self.wait_time = m.get('WaitTime')
|
|
168
|
+
return self
|
|
169
|
+
|
|
170
|
+
|
|
171
|
+
class AIAgentConfigAutoSpeechConfigUserIdleMessages(TeaModel):
|
|
172
|
+
def __init__(
|
|
173
|
+
self,
|
|
174
|
+
probability: float = None,
|
|
175
|
+
text: str = None,
|
|
176
|
+
):
|
|
177
|
+
self.probability = probability
|
|
178
|
+
self.text = text
|
|
179
|
+
|
|
180
|
+
def validate(self):
|
|
181
|
+
pass
|
|
182
|
+
|
|
183
|
+
def to_map(self):
|
|
184
|
+
_map = super().to_map()
|
|
185
|
+
if _map is not None:
|
|
186
|
+
return _map
|
|
187
|
+
|
|
188
|
+
result = dict()
|
|
189
|
+
if self.probability is not None:
|
|
190
|
+
result['Probability'] = self.probability
|
|
191
|
+
if self.text is not None:
|
|
192
|
+
result['Text'] = self.text
|
|
193
|
+
return result
|
|
194
|
+
|
|
195
|
+
def from_map(self, m: dict = None):
|
|
196
|
+
m = m or dict()
|
|
197
|
+
if m.get('Probability') is not None:
|
|
198
|
+
self.probability = m.get('Probability')
|
|
199
|
+
if m.get('Text') is not None:
|
|
200
|
+
self.text = m.get('Text')
|
|
201
|
+
return self
|
|
202
|
+
|
|
203
|
+
|
|
204
|
+
class AIAgentConfigAutoSpeechConfigUserIdle(TeaModel):
|
|
205
|
+
def __init__(
|
|
206
|
+
self,
|
|
207
|
+
max_repeats: int = None,
|
|
208
|
+
messages: List[AIAgentConfigAutoSpeechConfigUserIdleMessages] = None,
|
|
209
|
+
wait_time: int = None,
|
|
210
|
+
):
|
|
211
|
+
self.max_repeats = max_repeats
|
|
212
|
+
self.messages = messages
|
|
213
|
+
self.wait_time = wait_time
|
|
214
|
+
|
|
215
|
+
def validate(self):
|
|
216
|
+
if self.messages:
|
|
217
|
+
for k in self.messages:
|
|
218
|
+
if k:
|
|
219
|
+
k.validate()
|
|
220
|
+
|
|
221
|
+
def to_map(self):
|
|
222
|
+
_map = super().to_map()
|
|
223
|
+
if _map is not None:
|
|
224
|
+
return _map
|
|
225
|
+
|
|
226
|
+
result = dict()
|
|
227
|
+
if self.max_repeats is not None:
|
|
228
|
+
result['MaxRepeats'] = self.max_repeats
|
|
229
|
+
result['Messages'] = []
|
|
230
|
+
if self.messages is not None:
|
|
231
|
+
for k in self.messages:
|
|
232
|
+
result['Messages'].append(k.to_map() if k else None)
|
|
233
|
+
if self.wait_time is not None:
|
|
234
|
+
result['WaitTime'] = self.wait_time
|
|
235
|
+
return result
|
|
236
|
+
|
|
237
|
+
def from_map(self, m: dict = None):
|
|
238
|
+
m = m or dict()
|
|
239
|
+
if m.get('MaxRepeats') is not None:
|
|
240
|
+
self.max_repeats = m.get('MaxRepeats')
|
|
241
|
+
self.messages = []
|
|
242
|
+
if m.get('Messages') is not None:
|
|
243
|
+
for k in m.get('Messages'):
|
|
244
|
+
temp_model = AIAgentConfigAutoSpeechConfigUserIdleMessages()
|
|
245
|
+
self.messages.append(temp_model.from_map(k))
|
|
246
|
+
if m.get('WaitTime') is not None:
|
|
247
|
+
self.wait_time = m.get('WaitTime')
|
|
248
|
+
return self
|
|
249
|
+
|
|
250
|
+
|
|
251
|
+
class AIAgentConfigAutoSpeechConfig(TeaModel):
|
|
252
|
+
def __init__(
|
|
253
|
+
self,
|
|
254
|
+
llm_pending: AIAgentConfigAutoSpeechConfigLlmPending = None,
|
|
255
|
+
user_idle: AIAgentConfigAutoSpeechConfigUserIdle = None,
|
|
256
|
+
):
|
|
257
|
+
self.llm_pending = llm_pending
|
|
258
|
+
self.user_idle = user_idle
|
|
259
|
+
|
|
260
|
+
def validate(self):
|
|
261
|
+
if self.llm_pending:
|
|
262
|
+
self.llm_pending.validate()
|
|
263
|
+
if self.user_idle:
|
|
264
|
+
self.user_idle.validate()
|
|
265
|
+
|
|
266
|
+
def to_map(self):
|
|
267
|
+
_map = super().to_map()
|
|
268
|
+
if _map is not None:
|
|
269
|
+
return _map
|
|
270
|
+
|
|
271
|
+
result = dict()
|
|
272
|
+
if self.llm_pending is not None:
|
|
273
|
+
result['LlmPending'] = self.llm_pending.to_map()
|
|
274
|
+
if self.user_idle is not None:
|
|
275
|
+
result['UserIdle'] = self.user_idle.to_map()
|
|
276
|
+
return result
|
|
277
|
+
|
|
278
|
+
def from_map(self, m: dict = None):
|
|
279
|
+
m = m or dict()
|
|
280
|
+
if m.get('LlmPending') is not None:
|
|
281
|
+
temp_model = AIAgentConfigAutoSpeechConfigLlmPending()
|
|
282
|
+
self.llm_pending = temp_model.from_map(m['LlmPending'])
|
|
283
|
+
if m.get('UserIdle') is not None:
|
|
284
|
+
temp_model = AIAgentConfigAutoSpeechConfigUserIdle()
|
|
285
|
+
self.user_idle = temp_model.from_map(m['UserIdle'])
|
|
286
|
+
return self
|
|
287
|
+
|
|
288
|
+
|
|
97
289
|
class AIAgentConfigAvatarConfig(TeaModel):
|
|
98
290
|
def __init__(
|
|
99
291
|
self,
|
|
@@ -225,6 +417,7 @@ class AIAgentConfigLlmConfig(TeaModel):
|
|
|
225
417
|
self,
|
|
226
418
|
bailian_app_params: str = None,
|
|
227
419
|
function_map: List[AIAgentConfigLlmConfigFunctionMap] = None,
|
|
420
|
+
history_sync_with_tts: bool = None,
|
|
228
421
|
llm_complete_reply: bool = None,
|
|
229
422
|
llm_history: List[AIAgentConfigLlmConfigLlmHistory] = None,
|
|
230
423
|
llm_history_limit: int = None,
|
|
@@ -235,6 +428,7 @@ class AIAgentConfigLlmConfig(TeaModel):
|
|
|
235
428
|
):
|
|
236
429
|
self.bailian_app_params = bailian_app_params
|
|
237
430
|
self.function_map = function_map
|
|
431
|
+
self.history_sync_with_tts = history_sync_with_tts
|
|
238
432
|
self.llm_complete_reply = llm_complete_reply
|
|
239
433
|
self.llm_history = llm_history
|
|
240
434
|
self.llm_history_limit = llm_history_limit
|
|
@@ -265,6 +459,8 @@ class AIAgentConfigLlmConfig(TeaModel):
|
|
|
265
459
|
if self.function_map is not None:
|
|
266
460
|
for k in self.function_map:
|
|
267
461
|
result['FunctionMap'].append(k.to_map() if k else None)
|
|
462
|
+
if self.history_sync_with_tts is not None:
|
|
463
|
+
result['HistorySyncWithTTS'] = self.history_sync_with_tts
|
|
268
464
|
if self.llm_complete_reply is not None:
|
|
269
465
|
result['LlmCompleteReply'] = self.llm_complete_reply
|
|
270
466
|
result['LlmHistory'] = []
|
|
@@ -292,6 +488,8 @@ class AIAgentConfigLlmConfig(TeaModel):
|
|
|
292
488
|
for k in m.get('FunctionMap'):
|
|
293
489
|
temp_model = AIAgentConfigLlmConfigFunctionMap()
|
|
294
490
|
self.function_map.append(temp_model.from_map(k))
|
|
491
|
+
if m.get('HistorySyncWithTTS') is not None:
|
|
492
|
+
self.history_sync_with_tts = m.get('HistorySyncWithTTS')
|
|
295
493
|
if m.get('LlmCompleteReply') is not None:
|
|
296
494
|
self.llm_complete_reply = m.get('LlmCompleteReply')
|
|
297
495
|
self.llm_history = []
|
|
@@ -747,6 +945,7 @@ class AIAgentConfig(TeaModel):
|
|
|
747
945
|
self,
|
|
748
946
|
ambient_sound_config: AIAgentConfigAmbientSoundConfig = None,
|
|
749
947
|
asr_config: AIAgentConfigAsrConfig = None,
|
|
948
|
+
auto_speech_config: AIAgentConfigAutoSpeechConfig = None,
|
|
750
949
|
avatar_config: AIAgentConfigAvatarConfig = None,
|
|
751
950
|
avatar_url: str = None,
|
|
752
951
|
avatar_url_type: str = None,
|
|
@@ -770,6 +969,7 @@ class AIAgentConfig(TeaModel):
|
|
|
770
969
|
):
|
|
771
970
|
self.ambient_sound_config = ambient_sound_config
|
|
772
971
|
self.asr_config = asr_config
|
|
972
|
+
self.auto_speech_config = auto_speech_config
|
|
773
973
|
self.avatar_config = avatar_config
|
|
774
974
|
self.avatar_url = avatar_url
|
|
775
975
|
self.avatar_url_type = avatar_url_type
|
|
@@ -796,6 +996,8 @@ class AIAgentConfig(TeaModel):
|
|
|
796
996
|
self.ambient_sound_config.validate()
|
|
797
997
|
if self.asr_config:
|
|
798
998
|
self.asr_config.validate()
|
|
999
|
+
if self.auto_speech_config:
|
|
1000
|
+
self.auto_speech_config.validate()
|
|
799
1001
|
if self.avatar_config:
|
|
800
1002
|
self.avatar_config.validate()
|
|
801
1003
|
if self.interrupt_config:
|
|
@@ -821,6 +1023,8 @@ class AIAgentConfig(TeaModel):
|
|
|
821
1023
|
result['AmbientSoundConfig'] = self.ambient_sound_config.to_map()
|
|
822
1024
|
if self.asr_config is not None:
|
|
823
1025
|
result['AsrConfig'] = self.asr_config.to_map()
|
|
1026
|
+
if self.auto_speech_config is not None:
|
|
1027
|
+
result['AutoSpeechConfig'] = self.auto_speech_config.to_map()
|
|
824
1028
|
if self.avatar_config is not None:
|
|
825
1029
|
result['AvatarConfig'] = self.avatar_config.to_map()
|
|
826
1030
|
if self.avatar_url is not None:
|
|
@@ -871,6 +1075,9 @@ class AIAgentConfig(TeaModel):
|
|
|
871
1075
|
if m.get('AsrConfig') is not None:
|
|
872
1076
|
temp_model = AIAgentConfigAsrConfig()
|
|
873
1077
|
self.asr_config = temp_model.from_map(m['AsrConfig'])
|
|
1078
|
+
if m.get('AutoSpeechConfig') is not None:
|
|
1079
|
+
temp_model = AIAgentConfigAutoSpeechConfig()
|
|
1080
|
+
self.auto_speech_config = temp_model.from_map(m['AutoSpeechConfig'])
|
|
874
1081
|
if m.get('AvatarConfig') is not None:
|
|
875
1082
|
temp_model = AIAgentConfigAvatarConfig()
|
|
876
1083
|
self.avatar_config = temp_model.from_map(m['AvatarConfig'])
|
|
@@ -1011,6 +1218,198 @@ class AIAgentOutboundCallConfigAsrConfig(TeaModel):
|
|
|
1011
1218
|
return self
|
|
1012
1219
|
|
|
1013
1220
|
|
|
1221
|
+
class AIAgentOutboundCallConfigAutoSpeechConfigLlmPendingMessages(TeaModel):
|
|
1222
|
+
def __init__(
|
|
1223
|
+
self,
|
|
1224
|
+
probability: float = None,
|
|
1225
|
+
text: str = None,
|
|
1226
|
+
):
|
|
1227
|
+
self.probability = probability
|
|
1228
|
+
self.text = text
|
|
1229
|
+
|
|
1230
|
+
def validate(self):
|
|
1231
|
+
pass
|
|
1232
|
+
|
|
1233
|
+
def to_map(self):
|
|
1234
|
+
_map = super().to_map()
|
|
1235
|
+
if _map is not None:
|
|
1236
|
+
return _map
|
|
1237
|
+
|
|
1238
|
+
result = dict()
|
|
1239
|
+
if self.probability is not None:
|
|
1240
|
+
result['Probability'] = self.probability
|
|
1241
|
+
if self.text is not None:
|
|
1242
|
+
result['Text'] = self.text
|
|
1243
|
+
return result
|
|
1244
|
+
|
|
1245
|
+
def from_map(self, m: dict = None):
|
|
1246
|
+
m = m or dict()
|
|
1247
|
+
if m.get('Probability') is not None:
|
|
1248
|
+
self.probability = m.get('Probability')
|
|
1249
|
+
if m.get('Text') is not None:
|
|
1250
|
+
self.text = m.get('Text')
|
|
1251
|
+
return self
|
|
1252
|
+
|
|
1253
|
+
|
|
1254
|
+
class AIAgentOutboundCallConfigAutoSpeechConfigLlmPending(TeaModel):
|
|
1255
|
+
def __init__(
|
|
1256
|
+
self,
|
|
1257
|
+
messages: List[AIAgentOutboundCallConfigAutoSpeechConfigLlmPendingMessages] = None,
|
|
1258
|
+
wait_time: int = None,
|
|
1259
|
+
):
|
|
1260
|
+
self.messages = messages
|
|
1261
|
+
self.wait_time = wait_time
|
|
1262
|
+
|
|
1263
|
+
def validate(self):
|
|
1264
|
+
if self.messages:
|
|
1265
|
+
for k in self.messages:
|
|
1266
|
+
if k:
|
|
1267
|
+
k.validate()
|
|
1268
|
+
|
|
1269
|
+
def to_map(self):
|
|
1270
|
+
_map = super().to_map()
|
|
1271
|
+
if _map is not None:
|
|
1272
|
+
return _map
|
|
1273
|
+
|
|
1274
|
+
result = dict()
|
|
1275
|
+
result['Messages'] = []
|
|
1276
|
+
if self.messages is not None:
|
|
1277
|
+
for k in self.messages:
|
|
1278
|
+
result['Messages'].append(k.to_map() if k else None)
|
|
1279
|
+
if self.wait_time is not None:
|
|
1280
|
+
result['WaitTime'] = self.wait_time
|
|
1281
|
+
return result
|
|
1282
|
+
|
|
1283
|
+
def from_map(self, m: dict = None):
|
|
1284
|
+
m = m or dict()
|
|
1285
|
+
self.messages = []
|
|
1286
|
+
if m.get('Messages') is not None:
|
|
1287
|
+
for k in m.get('Messages'):
|
|
1288
|
+
temp_model = AIAgentOutboundCallConfigAutoSpeechConfigLlmPendingMessages()
|
|
1289
|
+
self.messages.append(temp_model.from_map(k))
|
|
1290
|
+
if m.get('WaitTime') is not None:
|
|
1291
|
+
self.wait_time = m.get('WaitTime')
|
|
1292
|
+
return self
|
|
1293
|
+
|
|
1294
|
+
|
|
1295
|
+
class AIAgentOutboundCallConfigAutoSpeechConfigUserIdleMessages(TeaModel):
|
|
1296
|
+
def __init__(
|
|
1297
|
+
self,
|
|
1298
|
+
probability: float = None,
|
|
1299
|
+
text: str = None,
|
|
1300
|
+
):
|
|
1301
|
+
self.probability = probability
|
|
1302
|
+
self.text = text
|
|
1303
|
+
|
|
1304
|
+
def validate(self):
|
|
1305
|
+
pass
|
|
1306
|
+
|
|
1307
|
+
def to_map(self):
|
|
1308
|
+
_map = super().to_map()
|
|
1309
|
+
if _map is not None:
|
|
1310
|
+
return _map
|
|
1311
|
+
|
|
1312
|
+
result = dict()
|
|
1313
|
+
if self.probability is not None:
|
|
1314
|
+
result['Probability'] = self.probability
|
|
1315
|
+
if self.text is not None:
|
|
1316
|
+
result['Text'] = self.text
|
|
1317
|
+
return result
|
|
1318
|
+
|
|
1319
|
+
def from_map(self, m: dict = None):
|
|
1320
|
+
m = m or dict()
|
|
1321
|
+
if m.get('Probability') is not None:
|
|
1322
|
+
self.probability = m.get('Probability')
|
|
1323
|
+
if m.get('Text') is not None:
|
|
1324
|
+
self.text = m.get('Text')
|
|
1325
|
+
return self
|
|
1326
|
+
|
|
1327
|
+
|
|
1328
|
+
class AIAgentOutboundCallConfigAutoSpeechConfigUserIdle(TeaModel):
|
|
1329
|
+
def __init__(
|
|
1330
|
+
self,
|
|
1331
|
+
max_repeats: int = None,
|
|
1332
|
+
messages: List[AIAgentOutboundCallConfigAutoSpeechConfigUserIdleMessages] = None,
|
|
1333
|
+
wait_time: int = None,
|
|
1334
|
+
):
|
|
1335
|
+
self.max_repeats = max_repeats
|
|
1336
|
+
self.messages = messages
|
|
1337
|
+
self.wait_time = wait_time
|
|
1338
|
+
|
|
1339
|
+
def validate(self):
|
|
1340
|
+
if self.messages:
|
|
1341
|
+
for k in self.messages:
|
|
1342
|
+
if k:
|
|
1343
|
+
k.validate()
|
|
1344
|
+
|
|
1345
|
+
def to_map(self):
|
|
1346
|
+
_map = super().to_map()
|
|
1347
|
+
if _map is not None:
|
|
1348
|
+
return _map
|
|
1349
|
+
|
|
1350
|
+
result = dict()
|
|
1351
|
+
if self.max_repeats is not None:
|
|
1352
|
+
result['MaxRepeats'] = self.max_repeats
|
|
1353
|
+
result['Messages'] = []
|
|
1354
|
+
if self.messages is not None:
|
|
1355
|
+
for k in self.messages:
|
|
1356
|
+
result['Messages'].append(k.to_map() if k else None)
|
|
1357
|
+
if self.wait_time is not None:
|
|
1358
|
+
result['WaitTime'] = self.wait_time
|
|
1359
|
+
return result
|
|
1360
|
+
|
|
1361
|
+
def from_map(self, m: dict = None):
|
|
1362
|
+
m = m or dict()
|
|
1363
|
+
if m.get('MaxRepeats') is not None:
|
|
1364
|
+
self.max_repeats = m.get('MaxRepeats')
|
|
1365
|
+
self.messages = []
|
|
1366
|
+
if m.get('Messages') is not None:
|
|
1367
|
+
for k in m.get('Messages'):
|
|
1368
|
+
temp_model = AIAgentOutboundCallConfigAutoSpeechConfigUserIdleMessages()
|
|
1369
|
+
self.messages.append(temp_model.from_map(k))
|
|
1370
|
+
if m.get('WaitTime') is not None:
|
|
1371
|
+
self.wait_time = m.get('WaitTime')
|
|
1372
|
+
return self
|
|
1373
|
+
|
|
1374
|
+
|
|
1375
|
+
class AIAgentOutboundCallConfigAutoSpeechConfig(TeaModel):
|
|
1376
|
+
def __init__(
|
|
1377
|
+
self,
|
|
1378
|
+
llm_pending: AIAgentOutboundCallConfigAutoSpeechConfigLlmPending = None,
|
|
1379
|
+
user_idle: AIAgentOutboundCallConfigAutoSpeechConfigUserIdle = None,
|
|
1380
|
+
):
|
|
1381
|
+
self.llm_pending = llm_pending
|
|
1382
|
+
self.user_idle = user_idle
|
|
1383
|
+
|
|
1384
|
+
def validate(self):
|
|
1385
|
+
if self.llm_pending:
|
|
1386
|
+
self.llm_pending.validate()
|
|
1387
|
+
if self.user_idle:
|
|
1388
|
+
self.user_idle.validate()
|
|
1389
|
+
|
|
1390
|
+
def to_map(self):
|
|
1391
|
+
_map = super().to_map()
|
|
1392
|
+
if _map is not None:
|
|
1393
|
+
return _map
|
|
1394
|
+
|
|
1395
|
+
result = dict()
|
|
1396
|
+
if self.llm_pending is not None:
|
|
1397
|
+
result['LlmPending'] = self.llm_pending.to_map()
|
|
1398
|
+
if self.user_idle is not None:
|
|
1399
|
+
result['UserIdle'] = self.user_idle.to_map()
|
|
1400
|
+
return result
|
|
1401
|
+
|
|
1402
|
+
def from_map(self, m: dict = None):
|
|
1403
|
+
m = m or dict()
|
|
1404
|
+
if m.get('LlmPending') is not None:
|
|
1405
|
+
temp_model = AIAgentOutboundCallConfigAutoSpeechConfigLlmPending()
|
|
1406
|
+
self.llm_pending = temp_model.from_map(m['LlmPending'])
|
|
1407
|
+
if m.get('UserIdle') is not None:
|
|
1408
|
+
temp_model = AIAgentOutboundCallConfigAutoSpeechConfigUserIdle()
|
|
1409
|
+
self.user_idle = temp_model.from_map(m['UserIdle'])
|
|
1410
|
+
return self
|
|
1411
|
+
|
|
1412
|
+
|
|
1014
1413
|
class AIAgentOutboundCallConfigInterruptConfig(TeaModel):
|
|
1015
1414
|
def __init__(
|
|
1016
1415
|
self,
|
|
@@ -1115,6 +1514,7 @@ class AIAgentOutboundCallConfigLlmConfig(TeaModel):
|
|
|
1115
1514
|
self,
|
|
1116
1515
|
bailian_app_params: str = None,
|
|
1117
1516
|
function_map: List[AIAgentOutboundCallConfigLlmConfigFunctionMap] = None,
|
|
1517
|
+
history_sync_with_tts: bool = None,
|
|
1118
1518
|
llm_complete_reply: bool = None,
|
|
1119
1519
|
llm_history: List[AIAgentOutboundCallConfigLlmConfigLlmHistory] = None,
|
|
1120
1520
|
llm_history_limit: int = None,
|
|
@@ -1125,6 +1525,7 @@ class AIAgentOutboundCallConfigLlmConfig(TeaModel):
|
|
|
1125
1525
|
):
|
|
1126
1526
|
self.bailian_app_params = bailian_app_params
|
|
1127
1527
|
self.function_map = function_map
|
|
1528
|
+
self.history_sync_with_tts = history_sync_with_tts
|
|
1128
1529
|
self.llm_complete_reply = llm_complete_reply
|
|
1129
1530
|
self.llm_history = llm_history
|
|
1130
1531
|
self.llm_history_limit = llm_history_limit
|
|
@@ -1155,6 +1556,8 @@ class AIAgentOutboundCallConfigLlmConfig(TeaModel):
|
|
|
1155
1556
|
if self.function_map is not None:
|
|
1156
1557
|
for k in self.function_map:
|
|
1157
1558
|
result['FunctionMap'].append(k.to_map() if k else None)
|
|
1559
|
+
if self.history_sync_with_tts is not None:
|
|
1560
|
+
result['HistorySyncWithTTS'] = self.history_sync_with_tts
|
|
1158
1561
|
if self.llm_complete_reply is not None:
|
|
1159
1562
|
result['LlmCompleteReply'] = self.llm_complete_reply
|
|
1160
1563
|
result['LlmHistory'] = []
|
|
@@ -1182,6 +1585,8 @@ class AIAgentOutboundCallConfigLlmConfig(TeaModel):
|
|
|
1182
1585
|
for k in m.get('FunctionMap'):
|
|
1183
1586
|
temp_model = AIAgentOutboundCallConfigLlmConfigFunctionMap()
|
|
1184
1587
|
self.function_map.append(temp_model.from_map(k))
|
|
1588
|
+
if m.get('HistorySyncWithTTS') is not None:
|
|
1589
|
+
self.history_sync_with_tts = m.get('HistorySyncWithTTS')
|
|
1185
1590
|
if m.get('LlmCompleteReply') is not None:
|
|
1186
1591
|
self.llm_complete_reply = m.get('LlmCompleteReply')
|
|
1187
1592
|
self.llm_history = []
|
|
@@ -1356,23 +1761,27 @@ class AIAgentOutboundCallConfig(TeaModel):
|
|
|
1356
1761
|
self,
|
|
1357
1762
|
ambient_sound_config: AIAgentOutboundCallConfigAmbientSoundConfig = None,
|
|
1358
1763
|
asr_config: AIAgentOutboundCallConfigAsrConfig = None,
|
|
1764
|
+
auto_speech_config: AIAgentOutboundCallConfigAutoSpeechConfig = None,
|
|
1359
1765
|
enable_intelligent_segment: bool = None,
|
|
1360
1766
|
experimental_config: str = None,
|
|
1361
1767
|
greeting: str = None,
|
|
1362
1768
|
greeting_delay: int = None,
|
|
1363
1769
|
interrupt_config: AIAgentOutboundCallConfigInterruptConfig = None,
|
|
1364
1770
|
llm_config: AIAgentOutboundCallConfigLlmConfig = None,
|
|
1771
|
+
max_idle_time: int = None,
|
|
1365
1772
|
tts_config: AIAgentOutboundCallConfigTtsConfig = None,
|
|
1366
1773
|
turn_detection_config: AIAgentOutboundCallConfigTurnDetectionConfig = None,
|
|
1367
1774
|
):
|
|
1368
1775
|
self.ambient_sound_config = ambient_sound_config
|
|
1369
1776
|
self.asr_config = asr_config
|
|
1777
|
+
self.auto_speech_config = auto_speech_config
|
|
1370
1778
|
self.enable_intelligent_segment = enable_intelligent_segment
|
|
1371
1779
|
self.experimental_config = experimental_config
|
|
1372
1780
|
self.greeting = greeting
|
|
1373
1781
|
self.greeting_delay = greeting_delay
|
|
1374
1782
|
self.interrupt_config = interrupt_config
|
|
1375
1783
|
self.llm_config = llm_config
|
|
1784
|
+
self.max_idle_time = max_idle_time
|
|
1376
1785
|
self.tts_config = tts_config
|
|
1377
1786
|
self.turn_detection_config = turn_detection_config
|
|
1378
1787
|
|
|
@@ -1381,6 +1790,8 @@ class AIAgentOutboundCallConfig(TeaModel):
|
|
|
1381
1790
|
self.ambient_sound_config.validate()
|
|
1382
1791
|
if self.asr_config:
|
|
1383
1792
|
self.asr_config.validate()
|
|
1793
|
+
if self.auto_speech_config:
|
|
1794
|
+
self.auto_speech_config.validate()
|
|
1384
1795
|
if self.interrupt_config:
|
|
1385
1796
|
self.interrupt_config.validate()
|
|
1386
1797
|
if self.llm_config:
|
|
@@ -1400,6 +1811,8 @@ class AIAgentOutboundCallConfig(TeaModel):
|
|
|
1400
1811
|
result['AmbientSoundConfig'] = self.ambient_sound_config.to_map()
|
|
1401
1812
|
if self.asr_config is not None:
|
|
1402
1813
|
result['AsrConfig'] = self.asr_config.to_map()
|
|
1814
|
+
if self.auto_speech_config is not None:
|
|
1815
|
+
result['AutoSpeechConfig'] = self.auto_speech_config.to_map()
|
|
1403
1816
|
if self.enable_intelligent_segment is not None:
|
|
1404
1817
|
result['EnableIntelligentSegment'] = self.enable_intelligent_segment
|
|
1405
1818
|
if self.experimental_config is not None:
|
|
@@ -1412,6 +1825,8 @@ class AIAgentOutboundCallConfig(TeaModel):
|
|
|
1412
1825
|
result['InterruptConfig'] = self.interrupt_config.to_map()
|
|
1413
1826
|
if self.llm_config is not None:
|
|
1414
1827
|
result['LlmConfig'] = self.llm_config.to_map()
|
|
1828
|
+
if self.max_idle_time is not None:
|
|
1829
|
+
result['MaxIdleTime'] = self.max_idle_time
|
|
1415
1830
|
if self.tts_config is not None:
|
|
1416
1831
|
result['TtsConfig'] = self.tts_config.to_map()
|
|
1417
1832
|
if self.turn_detection_config is not None:
|
|
@@ -1426,6 +1841,9 @@ class AIAgentOutboundCallConfig(TeaModel):
|
|
|
1426
1841
|
if m.get('AsrConfig') is not None:
|
|
1427
1842
|
temp_model = AIAgentOutboundCallConfigAsrConfig()
|
|
1428
1843
|
self.asr_config = temp_model.from_map(m['AsrConfig'])
|
|
1844
|
+
if m.get('AutoSpeechConfig') is not None:
|
|
1845
|
+
temp_model = AIAgentOutboundCallConfigAutoSpeechConfig()
|
|
1846
|
+
self.auto_speech_config = temp_model.from_map(m['AutoSpeechConfig'])
|
|
1429
1847
|
if m.get('EnableIntelligentSegment') is not None:
|
|
1430
1848
|
self.enable_intelligent_segment = m.get('EnableIntelligentSegment')
|
|
1431
1849
|
if m.get('ExperimentalConfig') is not None:
|
|
@@ -1440,6 +1858,8 @@ class AIAgentOutboundCallConfig(TeaModel):
|
|
|
1440
1858
|
if m.get('LlmConfig') is not None:
|
|
1441
1859
|
temp_model = AIAgentOutboundCallConfigLlmConfig()
|
|
1442
1860
|
self.llm_config = temp_model.from_map(m['LlmConfig'])
|
|
1861
|
+
if m.get('MaxIdleTime') is not None:
|
|
1862
|
+
self.max_idle_time = m.get('MaxIdleTime')
|
|
1443
1863
|
if m.get('TtsConfig') is not None:
|
|
1444
1864
|
temp_model = AIAgentOutboundCallConfigTtsConfig()
|
|
1445
1865
|
self.tts_config = temp_model.from_map(m['TtsConfig'])
|
|
@@ -26268,11 +26688,13 @@ class ForwardAIAgentCallRequest(TeaModel):
|
|
|
26268
26688
|
def __init__(
|
|
26269
26689
|
self,
|
|
26270
26690
|
called_number: str = None,
|
|
26691
|
+
caller_number: str = None,
|
|
26271
26692
|
error_prompt: str = None,
|
|
26272
26693
|
instance_id: str = None,
|
|
26273
26694
|
transfer_prompt: str = None,
|
|
26274
26695
|
):
|
|
26275
26696
|
self.called_number = called_number
|
|
26697
|
+
self.caller_number = caller_number
|
|
26276
26698
|
self.error_prompt = error_prompt
|
|
26277
26699
|
self.instance_id = instance_id
|
|
26278
26700
|
self.transfer_prompt = transfer_prompt
|
|
@@ -26288,6 +26710,8 @@ class ForwardAIAgentCallRequest(TeaModel):
|
|
|
26288
26710
|
result = dict()
|
|
26289
26711
|
if self.called_number is not None:
|
|
26290
26712
|
result['CalledNumber'] = self.called_number
|
|
26713
|
+
if self.caller_number is not None:
|
|
26714
|
+
result['CallerNumber'] = self.caller_number
|
|
26291
26715
|
if self.error_prompt is not None:
|
|
26292
26716
|
result['ErrorPrompt'] = self.error_prompt
|
|
26293
26717
|
if self.instance_id is not None:
|
|
@@ -26300,6 +26724,8 @@ class ForwardAIAgentCallRequest(TeaModel):
|
|
|
26300
26724
|
m = m or dict()
|
|
26301
26725
|
if m.get('CalledNumber') is not None:
|
|
26302
26726
|
self.called_number = m.get('CalledNumber')
|
|
26727
|
+
if m.get('CallerNumber') is not None:
|
|
26728
|
+
self.caller_number = m.get('CallerNumber')
|
|
26303
26729
|
if m.get('ErrorPrompt') is not None:
|
|
26304
26730
|
self.error_prompt = m.get('ErrorPrompt')
|
|
26305
26731
|
if m.get('InstanceId') is not None:
|
|
@@ -0,0 +1,8 @@
|
|
|
1
|
+
alibabacloud_ice20201109/__init__.py,sha256=cCX4fSFJnSHNtOBuqVN7iHzLprP8a9-rhtdtjk5w9_Y,21
|
|
2
|
+
alibabacloud_ice20201109/client.py,sha256=rye7Vx9KCV2mwRnH2b9hqTRWyHiL0_dWL_JD25bHIAw,1870900
|
|
3
|
+
alibabacloud_ice20201109/models.py,sha256=4uHONmkSBVOyxhKAd28Z804QDn38fF7JZ4_-if2Gb98,3751316
|
|
4
|
+
alibabacloud_ice20201109-6.8.4.dist-info/LICENSE,sha256=0CFItL6bHvxqS44T6vlLoW2R4Zaic304OO3WxN0oXF0,600
|
|
5
|
+
alibabacloud_ice20201109-6.8.4.dist-info/METADATA,sha256=1oMy-UUJs3-6zgzxU0qTHWwMKblI4HeoxMPJWVdMfg4,2312
|
|
6
|
+
alibabacloud_ice20201109-6.8.4.dist-info/WHEEL,sha256=2wepM1nk4DS4eFpYrW1TTqPcoGNfHhhO_i5m4cOimbo,92
|
|
7
|
+
alibabacloud_ice20201109-6.8.4.dist-info/top_level.txt,sha256=Tdq86hkGJfaKKyNwqChNx5I788aFI6e_iTUmQOYYPK8,25
|
|
8
|
+
alibabacloud_ice20201109-6.8.4.dist-info/RECORD,,
|
|
@@ -1,8 +0,0 @@
|
|
|
1
|
-
alibabacloud_ice20201109/__init__.py,sha256=VvJJe-O6H-MOyZxUgVFi9OmZgNGRH9Fe9g-HLafH-qw,21
|
|
2
|
-
alibabacloud_ice20201109/client.py,sha256=5veHxy6pH5kUqJy-0AzCcgIFAs9vH6FLro7VdX88f8Y,1870666
|
|
3
|
-
alibabacloud_ice20201109/models.py,sha256=KZEFBFsYPoyQrxvEaLqssapOP61q8tZA4WuGG4cRBN0,3737447
|
|
4
|
-
alibabacloud_ice20201109-6.8.3.dist-info/LICENSE,sha256=0CFItL6bHvxqS44T6vlLoW2R4Zaic304OO3WxN0oXF0,600
|
|
5
|
-
alibabacloud_ice20201109-6.8.3.dist-info/METADATA,sha256=M_7Cg-iVrYx2KmdAM0hs4szRBINuJSzlz84mmooCG7U,2312
|
|
6
|
-
alibabacloud_ice20201109-6.8.3.dist-info/WHEEL,sha256=2wepM1nk4DS4eFpYrW1TTqPcoGNfHhhO_i5m4cOimbo,92
|
|
7
|
-
alibabacloud_ice20201109-6.8.3.dist-info/top_level.txt,sha256=Tdq86hkGJfaKKyNwqChNx5I788aFI6e_iTUmQOYYPK8,25
|
|
8
|
-
alibabacloud_ice20201109-6.8.3.dist-info/RECORD,,
|
{alibabacloud_ice20201109-6.8.3.dist-info → alibabacloud_ice20201109-6.8.4.dist-info}/LICENSE
RENAMED
|
File without changes
|
|
File without changes
|
{alibabacloud_ice20201109-6.8.3.dist-info → alibabacloud_ice20201109-6.8.4.dist-info}/top_level.txt
RENAMED
|
File without changes
|