@genpact/genome.mfe.mgtwin-app 1.1.267-alpha → 1.1.269-alpha
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/{ArrowLeftCircleIcon-4e699249.mjs → ArrowLeftCircleIcon-f6041bde.mjs} +1 -1
- package/dist/{AssessmentAMA-7d9d9067.mjs → AssessmentAMA-9b2b0033.mjs} +4 -4
- package/dist/{AssessmentSimulation-2bfd1b5e.mjs → AssessmentSimulation-6466fc51.mjs} +205 -204
- package/dist/{EditSkillAdminData-1b8f4b5e.mjs → EditSkillAdminData-a9ccdeba.mjs} +3 -3
- package/dist/{EditSkillData-5d50895a.mjs → EditSkillData-a2915cd0.mjs} +3 -3
- package/dist/{HomeView-343755d3.mjs → HomeView-8adcee4f.mjs} +5 -5
- package/dist/{Loading-a9dda3c1.mjs → Loading-5ea7160e.mjs} +1 -1
- package/dist/{LoadingSimulation-e1c451ea.mjs → LoadingSimulation-27c136fb.mjs} +1 -1
- package/dist/{Modal-abe82066.mjs → Modal-cdd2991e.mjs} +1 -1
- package/dist/{PromptTesting-605c020e.mjs → PromptTesting-1a40feb5.mjs} +2 -2
- package/dist/{SkillAutomation-95a6630a.mjs → SkillAutomation-dc35824a.mjs} +3 -3
- package/dist/{SkillData-53c5ee0b.mjs → SkillData-0e630c17.mjs} +3 -3
- package/dist/{UnderMaintenance-f9ad0af9.mjs → UnderMaintenance-bd135520.mjs} +1 -1
- package/dist/{_commonjsHelpers-89b7dddd.mjs → _commonjsHelpers-e1f9766e.mjs} +1 -1
- package/dist/{api-91aec5f8.mjs → api-0bcaa8e3.mjs} +1 -1
- package/dist/{main-b1d63f30.mjs → main-fb55944a.mjs} +933 -933
- package/dist/mgtwin.es.min.js +1 -1
- package/dist/mgtwin.umd.min.js +8 -8
- package/dist/style.css +1 -1
- package/package.json +1 -1
|
@@ -1,8 +1,8 @@
|
|
|
1
|
-
import { _ as Cl, a as ip, L as yl } from "./LoadingSimulation-
|
|
2
|
-
import { _ as Pl, a as sp, Q as Ot, M as Aa } from "./Modal-
|
|
3
|
-
import { d as Pe, r as M, H as Dt, o as N, c as j, a as R, k as St, u as be, C as Bt, q as Ce, y as Lo, h as ge, s as op, t as Ve, b as k, f as _e, i as zn, I as Vi, F as xo, e as qo, w as Jo, J as Rl, p as Wn, g as Vn, G as Gi, z as El, K as wl, n as kt, m as ap, L as cp, l as Al } from "./main-
|
|
4
|
-
import { g as xn, p as gt } from "./api-
|
|
5
|
-
import { c as B, a as Il, r as _l } from "./_commonjsHelpers-
|
|
1
|
+
import { _ as Cl, a as ip, L as yl } from "./LoadingSimulation-27c136fb.mjs";
|
|
2
|
+
import { _ as Pl, a as sp, Q as Ot, M as Aa } from "./Modal-cdd2991e.mjs";
|
|
3
|
+
import { d as Pe, r as M, H as Dt, o as N, c as j, a as R, k as St, u as be, C as Bt, q as Ce, y as Lo, h as ge, s as op, t as Ve, b as k, f as _e, i as zn, I as Vi, F as xo, e as qo, w as Jo, J as Rl, p as Wn, g as Vn, G as Gi, z as El, K as wl, n as kt, m as ap, L as cp, l as Al } from "./main-fb55944a.mjs";
|
|
4
|
+
import { g as xn, p as gt } from "./api-0bcaa8e3.mjs";
|
|
5
|
+
import { c as B, a as Il, r as _l } from "./_commonjsHelpers-e1f9766e.mjs";
|
|
6
6
|
import { _ as Ft } from "./_plugin-vue_export-helper-dad06003.mjs";
|
|
7
7
|
import "./loadernew-ac153a51.mjs";
|
|
8
8
|
const Tl = { class: "flex h-14 w-full flex-1 items-center justify-between bg-[#021c42] md:pr-6 lg:pr-6" }, bl = ["href"], Ml = /* @__PURE__ */ R("img", {
|
|
@@ -6514,12 +6514,12 @@ function Kf() {
|
|
|
6514
6514
|
p.token != null && p.token !== "" && (y[p.headerName] = p.token);
|
|
6515
6515
|
let P = u.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Endpoint, "");
|
|
6516
6516
|
if (!P) {
|
|
6517
|
-
const I = i.ConnectionFactoryBase.getHostSuffix(v),
|
|
6518
|
-
P = `${
|
|
6517
|
+
const I = i.ConnectionFactoryBase.getHostSuffix(v), w = u.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Host, `wss://${v}.${n.BaseUrl}${I}`);
|
|
6518
|
+
P = `${w.endsWith("/") ? w : w + "/"}${S}${n.ApiKey}/${C}`;
|
|
6519
6519
|
}
|
|
6520
6520
|
this.setCommonUrlParams(u, f, P);
|
|
6521
|
-
const
|
|
6522
|
-
return Promise.resolve(new t.WebsocketConnection(P, f, y, new c.WebsocketMessageFormatter(), t.ProxyInfo.fromRecognizerConfig(u),
|
|
6521
|
+
const A = u.parameters.getProperty("SPEECH-EnableWebsocketCompression", "false") === "true";
|
|
6522
|
+
return Promise.resolve(new t.WebsocketConnection(P, f, y, new c.WebsocketMessageFormatter(), t.ProxyInfo.fromRecognizerConfig(u), A, d));
|
|
6523
6523
|
}
|
|
6524
6524
|
}
|
|
6525
6525
|
return Tr.DialogConnectionFactory = n, n.ApiKey = "api", n.BaseUrl = "convai.speech", Tr;
|
|
@@ -8393,10 +8393,10 @@ function Sg() {
|
|
|
8393
8393
|
if (h) {
|
|
8394
8394
|
f = p.parameters.getProperty(i.PropertyId.SpeechServiceConnection_Endpoint), f || (f = "wss://" + p.parameters.getProperty(i.PropertyId.SpeechServiceConnection_Host, "transcribe.{region}.cts.speech{hostSuffix}") + "{path}"), f = e.StringUtils.formatString(f, m);
|
|
8395
8395
|
const P = new URL(f);
|
|
8396
|
-
P.searchParams.forEach((I,
|
|
8397
|
-
S[
|
|
8398
|
-
}), new n.TranscriberConnectionFactory().setQueryParams(S, p, f), S[a.QueryParameterNames.CtsMeetingId] = m[a.QueryParameterNames.CtsMeetingId], S[a.QueryParameterNames.CtsDeviceId] = m[a.QueryParameterNames.CtsDeviceId], v.isHost || (S[a.QueryParameterNames.CtsIsParticipant] = ""), a.QueryParameterNames.Format in S || (S[a.QueryParameterNames.Format] = "simple"), P.searchParams.forEach((I,
|
|
8399
|
-
P.searchParams.set(
|
|
8396
|
+
P.searchParams.forEach((I, w) => {
|
|
8397
|
+
S[w] = I;
|
|
8398
|
+
}), new n.TranscriberConnectionFactory().setQueryParams(S, p, f), S[a.QueryParameterNames.CtsMeetingId] = m[a.QueryParameterNames.CtsMeetingId], S[a.QueryParameterNames.CtsDeviceId] = m[a.QueryParameterNames.CtsDeviceId], v.isHost || (S[a.QueryParameterNames.CtsIsParticipant] = ""), a.QueryParameterNames.Format in S || (S[a.QueryParameterNames.Format] = "simple"), P.searchParams.forEach((I, w) => {
|
|
8399
|
+
P.searchParams.set(w, S[w]), delete S[w];
|
|
8400
8400
|
}), f = P.toString();
|
|
8401
8401
|
} else {
|
|
8402
8402
|
const P = new n.TranslationConnectionFactory();
|
|
@@ -8518,8 +8518,8 @@ function Cg() {
|
|
|
8518
8518
|
for (const y of S) {
|
|
8519
8519
|
const P = this.privProperties.getProperty(y);
|
|
8520
8520
|
if (P) {
|
|
8521
|
-
const
|
|
8522
|
-
this.privSpeechTranslationConfig.setProperty(
|
|
8521
|
+
const A = typeof y == "string" ? y : a.PropertyId[y];
|
|
8522
|
+
this.privSpeechTranslationConfig.setProperty(A, P);
|
|
8523
8523
|
}
|
|
8524
8524
|
}
|
|
8525
8525
|
const C = JSON.parse(this.privProperties.getProperty(e.ServicePropertiesPropertyName, "{}"));
|
|
@@ -11425,19 +11425,19 @@ function E() {
|
|
|
11425
11425
|
Object.defineProperty(t, "TranslationRecognitionResult", { enumerable: !0, get: function() {
|
|
11426
11426
|
return P.TranslationRecognitionResult;
|
|
11427
11427
|
} });
|
|
11428
|
-
var
|
|
11428
|
+
var A = _s;
|
|
11429
11429
|
Object.defineProperty(t, "TranslationSynthesisResult", { enumerable: !0, get: function() {
|
|
11430
|
-
return
|
|
11430
|
+
return A.TranslationSynthesisResult;
|
|
11431
11431
|
} });
|
|
11432
11432
|
var I = Vp;
|
|
11433
11433
|
Object.defineProperty(t, "ResultReason", { enumerable: !0, get: function() {
|
|
11434
11434
|
return I.ResultReason;
|
|
11435
11435
|
} });
|
|
11436
|
-
var
|
|
11436
|
+
var w = Of();
|
|
11437
11437
|
Object.defineProperty(t, "SpeechConfig", { enumerable: !0, get: function() {
|
|
11438
|
-
return
|
|
11438
|
+
return w.SpeechConfig;
|
|
11439
11439
|
} }), Object.defineProperty(t, "SpeechConfigImpl", { enumerable: !0, get: function() {
|
|
11440
|
-
return
|
|
11440
|
+
return w.SpeechConfigImpl;
|
|
11441
11441
|
} });
|
|
11442
11442
|
var O = Df();
|
|
11443
11443
|
Object.defineProperty(t, "SpeechTranslationConfig", { enumerable: !0, get: function() {
|
|
@@ -12330,8 +12330,8 @@ function tm() {
|
|
|
12330
12330
|
throw new e.ArgumentNullError("recognizerConfig");
|
|
12331
12331
|
this.privEnableSpeakerId = y.isSpeakerDiarizationEnabled, this.privMustReportEndOfStream = !1, this.privAuthentication = f, this.privConnectionFactory = S, this.privAudioSource = C, this.privRecognizerConfig = y, this.privIsDisposed = !1, this.privRecognizer = P, this.privRequestSession = new i.RequestSession(this.privAudioSource.id()), this.privConnectionEvents = new e.EventSource(), this.privServiceEvents = new e.EventSource(), this.privDynamicGrammar = new i.DynamicGrammarBuilder(), this.privSpeechContext = new i.SpeechContext(this.privDynamicGrammar), this.privAgentConfig = new i.AgentConfig(), this.privRecognizerConfig.parameters.getProperty(r.PropertyId.WebWorkerLoadType, "on").toLowerCase() === "on" && typeof Blob < "u" && typeof Worker < "u" ? this.privSetTimeout = e.Timeout.setTimeout : (typeof window < "u" && (this.privSetTimeout = window.setTimeout.bind(window)), typeof globalThis < "u" && (this.privSetTimeout = globalThis.setTimeout.bind(globalThis))), this.connectionEvents.attach((I) => {
|
|
12332
12332
|
if (I.name === "ConnectionClosedEvent") {
|
|
12333
|
-
const
|
|
12334
|
-
(
|
|
12333
|
+
const w = I;
|
|
12334
|
+
(w.statusCode === 1003 || w.statusCode === 1007 || w.statusCode === 1002 || w.statusCode === 4e3 || this.privRequestSession.numConnectionAttempts > this.privRecognizerConfig.maxRetryCount) && this.cancelRecognitionLocal(r.CancellationReason.Error, w.statusCode === 1007 ? r.CancellationErrorCode.BadRequestParameters : r.CancellationErrorCode.ConnectionFailure, `${w.reason} websocket error code: ${w.statusCode}`);
|
|
12335
12335
|
}
|
|
12336
12336
|
}), this.privEnableSpeakerId && (this.privDiarizationSessionId = (0, e.createNoDashGuid)());
|
|
12337
12337
|
}
|
|
@@ -12349,14 +12349,14 @@ function tm() {
|
|
|
12349
12349
|
targetLanguages: S
|
|
12350
12350
|
}, y !== void 0 && (this.privSpeechContext.getContext().translation.category = y), C !== void 0) {
|
|
12351
12351
|
const I = {};
|
|
12352
|
-
for (const
|
|
12353
|
-
I[
|
|
12352
|
+
for (const w of S)
|
|
12353
|
+
I[w] = C;
|
|
12354
12354
|
this.privSpeechContext.getContext().synthesis = {
|
|
12355
12355
|
defaultVoices: I
|
|
12356
12356
|
};
|
|
12357
12357
|
}
|
|
12358
|
-
const
|
|
12359
|
-
|
|
12358
|
+
const A = this.privSpeechContext.getContext().phraseDetection || {};
|
|
12359
|
+
A.onSuccess = { action: v.NextAction.Translate }, A.onInterim = { action: v.NextAction.Translate }, this.privSpeechContext.getContext().phraseDetection = A;
|
|
12360
12360
|
}
|
|
12361
12361
|
}
|
|
12362
12362
|
setSpeechSegmentationTimeoutJson() {
|
|
@@ -12368,43 +12368,43 @@ function tm() {
|
|
|
12368
12368
|
let P = !1;
|
|
12369
12369
|
if (C !== void 0) {
|
|
12370
12370
|
P = !0;
|
|
12371
|
-
let
|
|
12371
|
+
let A = a.SegmentationMode.Normal;
|
|
12372
12372
|
switch (C.toLowerCase()) {
|
|
12373
12373
|
case "default":
|
|
12374
12374
|
break;
|
|
12375
12375
|
case "time":
|
|
12376
|
-
|
|
12376
|
+
A = a.SegmentationMode.Custom;
|
|
12377
12377
|
break;
|
|
12378
12378
|
case "semantic":
|
|
12379
|
-
|
|
12379
|
+
A = a.SegmentationMode.Semantic;
|
|
12380
12380
|
break;
|
|
12381
12381
|
}
|
|
12382
|
-
y.mode =
|
|
12382
|
+
y.mode = A;
|
|
12383
12383
|
}
|
|
12384
12384
|
if (f !== void 0) {
|
|
12385
12385
|
P = !0;
|
|
12386
|
-
const
|
|
12387
|
-
y.mode = a.SegmentationMode.Custom, y.segmentationSilenceTimeoutMs =
|
|
12386
|
+
const A = parseInt(f, 10);
|
|
12387
|
+
y.mode = a.SegmentationMode.Custom, y.segmentationSilenceTimeoutMs = A;
|
|
12388
12388
|
}
|
|
12389
12389
|
if (S !== void 0) {
|
|
12390
12390
|
P = !0;
|
|
12391
|
-
const
|
|
12392
|
-
y.mode = a.SegmentationMode.Custom, y.segmentationForcedTimeoutMs =
|
|
12391
|
+
const A = parseInt(S, 10);
|
|
12392
|
+
y.mode = a.SegmentationMode.Custom, y.segmentationForcedTimeoutMs = A;
|
|
12393
12393
|
}
|
|
12394
12394
|
if (P) {
|
|
12395
|
-
const
|
|
12396
|
-
switch (
|
|
12395
|
+
const A = this.privSpeechContext.getContext().phraseDetection || {};
|
|
12396
|
+
switch (A.mode = this.recognitionMode, this.recognitionMode) {
|
|
12397
12397
|
case o.RecognitionMode.Conversation:
|
|
12398
|
-
|
|
12398
|
+
A.conversation = A.conversation ?? { segmentation: {} }, A.conversation.segmentation = y;
|
|
12399
12399
|
break;
|
|
12400
12400
|
case o.RecognitionMode.Interactive:
|
|
12401
|
-
|
|
12401
|
+
A.interactive = A.interactive ?? { segmentation: {} }, A.interactive.segmentation = y;
|
|
12402
12402
|
break;
|
|
12403
12403
|
case o.RecognitionMode.Dictation:
|
|
12404
|
-
|
|
12404
|
+
A.dictation = A.dictation ?? {}, A.dictation.segmentation = y;
|
|
12405
12405
|
break;
|
|
12406
12406
|
}
|
|
12407
|
-
this.privSpeechContext.getContext().phraseDetection =
|
|
12407
|
+
this.privSpeechContext.getContext().phraseDetection = A;
|
|
12408
12408
|
}
|
|
12409
12409
|
}
|
|
12410
12410
|
setLanguageIdJson() {
|
|
@@ -12485,27 +12485,27 @@ function tm() {
|
|
|
12485
12485
|
return;
|
|
12486
12486
|
}
|
|
12487
12487
|
if (this.privConnectionConfigurationPromise = void 0, this.privRecognizerConfig.recognitionMode = f, this.privRecognizerConfig.recognitionEndpointVersion === "2") {
|
|
12488
|
-
const
|
|
12489
|
-
|
|
12488
|
+
const w = this.privSpeechContext.getContext().phraseDetection || {};
|
|
12489
|
+
w.mode = f, this.privSpeechContext.getContext().phraseDetection = w;
|
|
12490
12490
|
}
|
|
12491
12491
|
this.setLanguageIdJson(), this.setTranslationJson(), this.privRecognizerConfig.autoDetectSourceLanguages !== void 0 && this.privRecognizerConfig.parameters.getProperty(r.PropertyId.SpeechServiceConnection_TranslationToLanguages, void 0) !== void 0 && this.setupTranslationWithLanguageId(), this.setSpeechSegmentationTimeoutJson(), this.setOutputDetailLevelJson(), this.privSuccessCallback = S, this.privErrorCallback = C, this.privRequestSession.startNewRecognition(), this.privRequestSession.listenForServiceTelemetry(this.privAudioSource.events);
|
|
12492
12492
|
const y = this.connectImpl();
|
|
12493
12493
|
let P;
|
|
12494
12494
|
try {
|
|
12495
|
-
const
|
|
12496
|
-
this.privIsLiveAudio = F.type && F.type === i.type.Microphones, P = new t.ReplayableAudioNode(
|
|
12497
|
-
} catch (
|
|
12498
|
-
throw await this.privRequestSession.onStopRecognizing(),
|
|
12495
|
+
const w = await this.audioSource.attach(this.privRequestSession.audioNodeId), O = await this.audioSource.format, F = await this.audioSource.deviceInfo;
|
|
12496
|
+
this.privIsLiveAudio = F.type && F.type === i.type.Microphones, P = new t.ReplayableAudioNode(w, O.avgBytesPerSec), await this.privRequestSession.onAudioSourceAttachCompleted(P, !1), this.privRecognizerConfig.SpeechServiceConfig.Context.audio = { source: F };
|
|
12497
|
+
} catch (w) {
|
|
12498
|
+
throw await this.privRequestSession.onStopRecognizing(), w;
|
|
12499
12499
|
}
|
|
12500
12500
|
try {
|
|
12501
12501
|
await y;
|
|
12502
|
-
} catch (
|
|
12503
|
-
await this.cancelRecognitionLocal(r.CancellationReason.Error, r.CancellationErrorCode.ConnectionFailure,
|
|
12502
|
+
} catch (w) {
|
|
12503
|
+
await this.cancelRecognitionLocal(r.CancellationReason.Error, r.CancellationErrorCode.ConnectionFailure, w);
|
|
12504
12504
|
return;
|
|
12505
12505
|
}
|
|
12506
|
-
const
|
|
12507
|
-
this.privRecognizer.sessionStarted && this.privRecognizer.sessionStarted(this.privRecognizer,
|
|
12508
|
-
await this.cancelRecognitionLocal(r.CancellationReason.Error, r.CancellationErrorCode.RuntimeError,
|
|
12506
|
+
const A = new r.SessionEventArgs(this.privRequestSession.sessionId);
|
|
12507
|
+
this.privRecognizer.sessionStarted && this.privRecognizer.sessionStarted(this.privRecognizer, A), this.receiveMessage(), this.sendAudio(P).catch(async (w) => {
|
|
12508
|
+
await this.cancelRecognitionLocal(r.CancellationReason.Error, r.CancellationErrorCode.RuntimeError, w);
|
|
12509
12509
|
});
|
|
12510
12510
|
}
|
|
12511
12511
|
async stopRecognizing() {
|
|
@@ -12594,10 +12594,10 @@ function tm() {
|
|
|
12594
12594
|
this.privRecognizer.speechStartDetected && this.privRecognizer.speechStartDetected(this.privRecognizer, P);
|
|
12595
12595
|
break;
|
|
12596
12596
|
case "speech.enddetected":
|
|
12597
|
-
let
|
|
12598
|
-
C.textBody.length > 0 ?
|
|
12599
|
-
const I = i.SpeechDetected.fromJSON(
|
|
12600
|
-
this.privRecognizer.speechEndDetected && this.privRecognizer.speechEndDetected(this.privRecognizer,
|
|
12597
|
+
let A;
|
|
12598
|
+
C.textBody.length > 0 ? A = C.textBody : A = "{ Offset: 0 }";
|
|
12599
|
+
const I = i.SpeechDetected.fromJSON(A, this.privRequestSession.currentTurnAudioOffset), w = new r.RecognitionEventArgs(I.Offset + this.privRequestSession.currentTurnAudioOffset, this.privRequestSession.sessionId);
|
|
12600
|
+
this.privRecognizer.speechEndDetected && this.privRecognizer.speechEndDetected(this.privRecognizer, w);
|
|
12601
12601
|
break;
|
|
12602
12602
|
case "turn.end":
|
|
12603
12603
|
await this.sendTelemetryData(), this.privRequestSession.isSpeechEnded && this.privMustReportEndOfStream && (this.privMustReportEndOfStream = !1, await this.cancelRecognitionLocal(r.CancellationReason.EndOfStream, r.CancellationErrorCode.NoError, void 0));
|
|
@@ -12641,8 +12641,8 @@ function tm() {
|
|
|
12641
12641
|
if (C) {
|
|
12642
12642
|
const y = this.privRecognizerConfig.sourceLanguageModels;
|
|
12643
12643
|
if (y !== void 0 && y.length > 0) {
|
|
12644
|
-
const
|
|
12645
|
-
|
|
12644
|
+
const A = this.privSpeechContext.getContext().phraseDetection || {};
|
|
12645
|
+
A.customModels = y, this.privSpeechContext.getContext().phraseDetection = A;
|
|
12646
12646
|
}
|
|
12647
12647
|
this.privRecognizerConfig.parameters.getProperty(r.PropertyId.SpeechServiceConnection_TranslationVoice, void 0) !== void 0 && (C.onSuccess = { action: n.NextAction.Synthesize }, C.onPassthrough = { action: n.NextAction.Synthesize });
|
|
12648
12648
|
}
|
|
@@ -12688,14 +12688,14 @@ function tm() {
|
|
|
12688
12688
|
const S = await this.audioSource.format;
|
|
12689
12689
|
this.privAverageBytesPerMs = S.avgBytesPerSec / 1e3;
|
|
12690
12690
|
let C = Date.now();
|
|
12691
|
-
const y = this.privRecognizerConfig.parameters.getProperty("SPEECH-TransmitLengthBeforThrottleMs", "5000"), P = S.avgBytesPerSec / 1e3 * parseInt(y, 10),
|
|
12692
|
-
if (!this.privIsDisposed && !this.privRequestSession.isSpeechEnded && this.privRequestSession.isRecognizing && this.privRequestSession.recogNumber ===
|
|
12693
|
-
const
|
|
12691
|
+
const y = this.privRecognizerConfig.parameters.getProperty("SPEECH-TransmitLengthBeforThrottleMs", "5000"), P = S.avgBytesPerSec / 1e3 * parseInt(y, 10), A = this.privRequestSession.recogNumber, I = async () => {
|
|
12692
|
+
if (!this.privIsDisposed && !this.privRequestSession.isSpeechEnded && this.privRequestSession.isRecognizing && this.privRequestSession.recogNumber === A) {
|
|
12693
|
+
const w = await this.fetchConnection(), O = await f.read();
|
|
12694
12694
|
if (this.privRequestSession.isSpeechEnded)
|
|
12695
12695
|
return;
|
|
12696
12696
|
let F, Q;
|
|
12697
|
-
if (!O || O.isEnd ? (F = null, Q = 0) : (F = O.buffer, this.privRequestSession.onAudioSent(F.byteLength), P >= this.privRequestSession.bytesSent ? Q = 0 : Q = Math.max(0, C - Date.now())), Q !== 0 && await this.delay(Q), F !== null && (C = Date.now() + F.byteLength * 1e3 / (S.avgBytesPerSec * 2)), !this.privIsDisposed && !this.privRequestSession.isSpeechEnded && this.privRequestSession.isRecognizing && this.privRequestSession.recogNumber ===
|
|
12698
|
-
if (
|
|
12697
|
+
if (!O || O.isEnd ? (F = null, Q = 0) : (F = O.buffer, this.privRequestSession.onAudioSent(F.byteLength), P >= this.privRequestSession.bytesSent ? Q = 0 : Q = Math.max(0, C - Date.now())), Q !== 0 && await this.delay(Q), F !== null && (C = Date.now() + F.byteLength * 1e3 / (S.avgBytesPerSec * 2)), !this.privIsDisposed && !this.privRequestSession.isSpeechEnded && this.privRequestSession.isRecognizing && this.privRequestSession.recogNumber === A)
|
|
12698
|
+
if (w.send(new c.SpeechConnectionMessage(e.MessageType.Binary, "audio", this.privRequestSession.requestId, null, F)).catch(() => {
|
|
12699
12699
|
this.privRequestSession.onServiceTurnEndResponse(this.privRecognizerConfig.isContinuousRecognition).catch(() => {
|
|
12700
12700
|
});
|
|
12701
12701
|
}), O != null && O.isEnd)
|
|
@@ -12714,16 +12714,16 @@ function tm() {
|
|
|
12714
12714
|
let C = 0, y = "";
|
|
12715
12715
|
for (; this.privRequestSession.numConnectionAttempts <= this.privRecognizerConfig.maxRetryCount; ) {
|
|
12716
12716
|
this.privRequestSession.onRetryConnection();
|
|
12717
|
-
const
|
|
12717
|
+
const A = await (f ? this.privAuthentication.fetchOnExpiry(this.privAuthFetchEventId) : this.privAuthentication.fetch(this.privAuthFetchEventId));
|
|
12718
12718
|
await this.privRequestSession.onAuthCompleted(!1);
|
|
12719
|
-
const I = await this.privConnectionFactory.create(this.privRecognizerConfig,
|
|
12719
|
+
const I = await this.privConnectionFactory.create(this.privRecognizerConfig, A, this.privConnectionId);
|
|
12720
12720
|
this.privRequestSession.listenForServiceTelemetry(I.events), I.events.attach((O) => {
|
|
12721
12721
|
this.connectionEvents.onEvent(O);
|
|
12722
12722
|
});
|
|
12723
|
-
const
|
|
12724
|
-
if (
|
|
12725
|
-
return await this.privRequestSession.onConnectionEstablishCompleted(
|
|
12726
|
-
|
|
12723
|
+
const w = await I.open();
|
|
12724
|
+
if (w.statusCode === 200)
|
|
12725
|
+
return await this.privRequestSession.onConnectionEstablishCompleted(w.statusCode), Promise.resolve(I);
|
|
12726
|
+
w.statusCode === 1006 && (f = !0), C = w.statusCode, y = w.reason;
|
|
12727
12727
|
}
|
|
12728
12728
|
return await this.privRequestSession.onConnectionEstablishCompleted(C, y), Promise.reject(`Unable to contact server. StatusCode: ${C}, ${this.privRecognizerConfig.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Endpoint)} Reason: ${y}`);
|
|
12729
12729
|
}
|
|
@@ -12978,8 +12978,8 @@ function sm() {
|
|
|
12978
12978
|
let h = p.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Endpoint, void 0);
|
|
12979
12979
|
const v = p.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Region, void 0), g = i.ConnectionFactoryBase.getHostSuffix(v), m = p.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Host, "wss://" + v + ".stt.speech" + g), f = {}, S = p.parameters.getProperty(r.PropertyId.SpeechServiceConnection_EndpointId, void 0), C = p.parameters.getProperty(r.PropertyId.SpeechServiceConnection_RecoLanguage, void 0);
|
|
12980
12980
|
if (S ? (!h || h.search(o.QueryParameterNames.CustomSpeechDeploymentId) === -1) && (f[o.QueryParameterNames.CustomSpeechDeploymentId] = S) : C && (!h || h.search(o.QueryParameterNames.Language) === -1) && (f[o.QueryParameterNames.Language] = C), (!h || h.search(o.QueryParameterNames.Format) === -1) && (f[o.QueryParameterNames.Format] = p.parameters.getProperty(e.OutputFormatPropertyName, r.OutputFormat[r.OutputFormat.Simple]).toLowerCase()), p.autoDetectSourceLanguages !== void 0 && (f[o.QueryParameterNames.EnableLanguageId] = "true"), this.setCommonUrlParams(p, f, h), h) {
|
|
12981
|
-
const
|
|
12982
|
-
(O === "" || O === "/") && (
|
|
12981
|
+
const w = new URL(h), O = w.pathname;
|
|
12982
|
+
(O === "" || O === "/") && (w.pathname = this.universalUri + p.recognitionEndpointVersion, h = await i.ConnectionFactoryBase.getRedirectUrlFromEndpoint(w.toString()));
|
|
12983
12983
|
}
|
|
12984
12984
|
if (!h)
|
|
12985
12985
|
switch (p.recognitionMode) {
|
|
@@ -12995,8 +12995,8 @@ function sm() {
|
|
|
12995
12995
|
}
|
|
12996
12996
|
const y = {};
|
|
12997
12997
|
d.token !== void 0 && d.token !== "" && (y[d.headerName] = d.token), y[a.HeaderNames.ConnectionId] = l, y.connectionId = l;
|
|
12998
|
-
const P = p.parameters.getProperty("SPEECH-EnableWebsocketCompression", "false") === "true",
|
|
12999
|
-
return p.parameters.setProperty(r.PropertyId.SpeechServiceConnection_Url, I),
|
|
12998
|
+
const P = p.parameters.getProperty("SPEECH-EnableWebsocketCompression", "false") === "true", A = new t.WebsocketConnection(h, f, y, new c.WebsocketMessageFormatter(), t.ProxyInfo.fromRecognizerConfig(p), P, l), I = A.uri;
|
|
12999
|
+
return p.parameters.setProperty(r.PropertyId.SpeechServiceConnection_Url, I), A;
|
|
13000
13000
|
}
|
|
13001
13001
|
};
|
|
13002
13002
|
return an.SpeechConnectionFactory = s, an;
|
|
@@ -13015,14 +13015,14 @@ function om() {
|
|
|
13015
13015
|
let l = u.parameters.getProperty(e.PropertyId.SpeechServiceConnection_Endpoint, void 0);
|
|
13016
13016
|
const h = u.parameters.getProperty(e.PropertyId.SpeechServiceConnection_Region, void 0), v = i.ConnectionFactoryBase.getHostSuffix(h), g = u.parameters.getProperty(e.PropertyId.SpeechServiceConnection_Host, "wss://" + h + ".stt.speech" + v), m = {}, f = u.parameters.getProperty(e.PropertyId.SpeechServiceConnection_EndpointId, void 0), S = u.parameters.getProperty(e.PropertyId.SpeechServiceConnection_RecoLanguage, void 0);
|
|
13017
13017
|
if (f ? (!l || l.search(o.QueryParameterNames.CustomSpeechDeploymentId) === -1) && (m[o.QueryParameterNames.CustomSpeechDeploymentId] = f) : S && (!l || l.search(o.QueryParameterNames.Language) === -1) && (m[o.QueryParameterNames.Language] = S), u.autoDetectSourceLanguages !== void 0 && (m[o.QueryParameterNames.EnableLanguageId] = "true"), this.setV2UrlParams(u, m, l), l) {
|
|
13018
|
-
const I = new URL(l),
|
|
13019
|
-
(
|
|
13018
|
+
const I = new URL(l), w = I.pathname;
|
|
13019
|
+
(w === "" || w === "/") && (I.pathname = this.universalUri, l = await i.ConnectionFactoryBase.getRedirectUrlFromEndpoint(I.toString()));
|
|
13020
13020
|
}
|
|
13021
13021
|
l || (l = `${g}${this.universalUri}`);
|
|
13022
13022
|
const C = {};
|
|
13023
13023
|
p.token !== void 0 && p.token !== "" && (C[p.headerName] = p.token), C[a.HeaderNames.ConnectionId] = d;
|
|
13024
|
-
const y = u.parameters.getProperty("SPEECH-EnableWebsocketCompression", "false") === "true", P = new t.WebsocketConnection(l, m, C, new c.WebsocketMessageFormatter(), t.ProxyInfo.fromRecognizerConfig(u), y, d),
|
|
13025
|
-
return u.parameters.setProperty(e.PropertyId.SpeechServiceConnection_Url,
|
|
13024
|
+
const y = u.parameters.getProperty("SPEECH-EnableWebsocketCompression", "false") === "true", P = new t.WebsocketConnection(l, m, C, new c.WebsocketMessageFormatter(), t.ProxyInfo.fromRecognizerConfig(u), y, d), A = P.uri;
|
|
13025
|
+
return u.parameters.setProperty(e.PropertyId.SpeechServiceConnection_Url, A), P;
|
|
13026
13026
|
}
|
|
13027
13027
|
setV2UrlParams(u, p, d) {
|
|
13028
13028
|
(/* @__PURE__ */ new Map([
|
|
@@ -14704,19 +14704,19 @@ function Nm() {
|
|
|
14704
14704
|
case "speech.keyword":
|
|
14705
14705
|
const P = o.SpeechKeyword.fromJSON(h.textBody, this.privRequestSession.currentTurnAudioOffset);
|
|
14706
14706
|
g = new c.SpeechRecognitionResult(this.privRequestSession.requestId, P.Status === "Accepted" ? c.ResultReason.RecognizedKeyword : c.ResultReason.NoMatch, P.Text, P.Duration, P.Offset, void 0, void 0, void 0, void 0, P.asJson(), v), P.Status !== "Accepted" && (this.privLastResult = g);
|
|
14707
|
-
const
|
|
14707
|
+
const A = new c.SpeechRecognitionEventArgs(g, g.duration, g.resultId);
|
|
14708
14708
|
if (this.privDialogServiceConnector.recognized)
|
|
14709
14709
|
try {
|
|
14710
|
-
this.privDialogServiceConnector.recognized(this.privDialogServiceConnector,
|
|
14710
|
+
this.privDialogServiceConnector.recognized(this.privDialogServiceConnector, A);
|
|
14711
14711
|
} catch {
|
|
14712
14712
|
}
|
|
14713
14713
|
m = !0;
|
|
14714
14714
|
break;
|
|
14715
14715
|
case "audio":
|
|
14716
14716
|
{
|
|
14717
|
-
const I = h.requestId.toUpperCase(),
|
|
14717
|
+
const I = h.requestId.toUpperCase(), w = this.privTurnStateManager.GetTurn(I);
|
|
14718
14718
|
try {
|
|
14719
|
-
h.binaryBody ?
|
|
14719
|
+
h.binaryBody ? w.audioStream.write(h.binaryBody) : w.endAudioStream();
|
|
14720
14720
|
} catch {
|
|
14721
14721
|
}
|
|
14722
14722
|
}
|
|
@@ -14774,13 +14774,13 @@ function Nm() {
|
|
|
14774
14774
|
await this.privRequestSession.onAudioSourceAttachCompleted(P, !1), this.privRecognizerConfig.SpeechServiceConfig.Context.audio = { source: y };
|
|
14775
14775
|
try {
|
|
14776
14776
|
await m, await f;
|
|
14777
|
-
} catch (
|
|
14778
|
-
return await this.cancelRecognition(this.privRequestSession.sessionId, this.privRequestSession.requestId, c.CancellationReason.Error, c.CancellationErrorCode.ConnectionFailure,
|
|
14777
|
+
} catch (w) {
|
|
14778
|
+
return await this.cancelRecognition(this.privRequestSession.sessionId, this.privRequestSession.requestId, c.CancellationReason.Error, c.CancellationErrorCode.ConnectionFailure, w), Promise.resolve();
|
|
14779
14779
|
}
|
|
14780
|
-
const
|
|
14781
|
-
this.privRecognizer.sessionStarted && this.privRecognizer.sessionStarted(this.privRecognizer,
|
|
14782
|
-
}, async (
|
|
14783
|
-
await this.cancelRecognition(this.privRequestSession.sessionId, this.privRequestSession.requestId, c.CancellationReason.Error, c.CancellationErrorCode.RuntimeError,
|
|
14780
|
+
const A = new c.SessionEventArgs(this.privRequestSession.sessionId);
|
|
14781
|
+
this.privRecognizer.sessionStarted && this.privRecognizer.sessionStarted(this.privRecognizer, A), this.sendAudio(P).then(() => {
|
|
14782
|
+
}, async (w) => {
|
|
14783
|
+
await this.cancelRecognition(this.privRequestSession.sessionId, this.privRequestSession.requestId, c.CancellationReason.Error, c.CancellationErrorCode.RuntimeError, w);
|
|
14784
14784
|
});
|
|
14785
14785
|
}
|
|
14786
14786
|
// Establishes a websocket connection to the end point.
|
|
@@ -14807,15 +14807,15 @@ function Nm() {
|
|
|
14807
14807
|
}
|
|
14808
14808
|
break;
|
|
14809
14809
|
case "speech.startdetected":
|
|
14810
|
-
const P = o.SpeechDetected.fromJSON(C.textBody, this.privRequestSession.currentTurnAudioOffset),
|
|
14811
|
-
this.privRecognizer.speechStartDetected && this.privRecognizer.speechStartDetected(this.privRecognizer,
|
|
14810
|
+
const P = o.SpeechDetected.fromJSON(C.textBody, this.privRequestSession.currentTurnAudioOffset), A = new c.RecognitionEventArgs(P.Offset, this.privRequestSession.sessionId);
|
|
14811
|
+
this.privRecognizer.speechStartDetected && this.privRecognizer.speechStartDetected(this.privRecognizer, A);
|
|
14812
14812
|
break;
|
|
14813
14813
|
case "speech.enddetected":
|
|
14814
14814
|
let I;
|
|
14815
14815
|
C.textBody.length > 0 ? I = C.textBody : I = "{ Offset: 0 }";
|
|
14816
|
-
const
|
|
14817
|
-
this.privRequestSession.onServiceRecognized(
|
|
14818
|
-
const O = new c.RecognitionEventArgs(
|
|
14816
|
+
const w = o.SpeechDetected.fromJSON(I, this.privRequestSession.currentTurnAudioOffset);
|
|
14817
|
+
this.privRequestSession.onServiceRecognized(w.Offset);
|
|
14818
|
+
const O = new c.RecognitionEventArgs(w.Offset, this.privRequestSession.sessionId);
|
|
14819
14819
|
this.privRecognizer.speechEndDetected && this.privRecognizer.speechEndDetected(this.privRecognizer, O);
|
|
14820
14820
|
break;
|
|
14821
14821
|
case "turn.end":
|
|
@@ -15029,25 +15029,25 @@ function Lm() {
|
|
|
15029
15029
|
C.headers = S, this.privRestAdapter.options = C;
|
|
15030
15030
|
const y = `https://${l}${this.privRestPath}`;
|
|
15031
15031
|
this.privRestAdapter.request(t.RestRequestType.Post, y, f, null).then((P) => {
|
|
15032
|
-
const
|
|
15032
|
+
const A = t.RestMessageAdapter.extractHeaderValue(this.privRequestParams.requestId, P.headers);
|
|
15033
15033
|
if (!P.ok) {
|
|
15034
15034
|
if (u) {
|
|
15035
|
-
let
|
|
15035
|
+
let w = this.privErrors.invalidCreateJoinConversationResponse.replace("{status}", P.status.toString()), O;
|
|
15036
15036
|
try {
|
|
15037
|
-
O = JSON.parse(P.data),
|
|
15037
|
+
O = JSON.parse(P.data), w += ` [${O.error.code}: ${O.error.message}]`;
|
|
15038
15038
|
} catch {
|
|
15039
|
-
|
|
15039
|
+
w += ` [${P.data}]`;
|
|
15040
15040
|
}
|
|
15041
|
-
|
|
15041
|
+
A && (w += ` ${A}`), u(w);
|
|
15042
15042
|
}
|
|
15043
15043
|
return;
|
|
15044
15044
|
}
|
|
15045
15045
|
const I = JSON.parse(P.data);
|
|
15046
|
-
if (I && (I.requestId =
|
|
15046
|
+
if (I && (I.requestId = A), s) {
|
|
15047
15047
|
try {
|
|
15048
15048
|
s(I);
|
|
15049
|
-
} catch (
|
|
15050
|
-
u && u(
|
|
15049
|
+
} catch (w) {
|
|
15050
|
+
u && u(w);
|
|
15051
15051
|
}
|
|
15052
15052
|
s = void 0;
|
|
15053
15053
|
}
|
|
@@ -15655,7 +15655,7 @@ function Gm() {
|
|
|
15655
15655
|
const m = n.CommandResponsePayload.fromJSON(l.textBody);
|
|
15656
15656
|
switch (m.command.toLowerCase()) {
|
|
15657
15657
|
case "participantlist":
|
|
15658
|
-
const
|
|
15658
|
+
const A = n.ParticipantsListPayloadResponse.fromJSON(l.textBody), I = A.participants.map((K) => ({
|
|
15659
15659
|
avatar: K.avatar,
|
|
15660
15660
|
displayName: K.nickname,
|
|
15661
15661
|
id: K.participantId,
|
|
@@ -15664,7 +15664,7 @@ function Gm() {
|
|
|
15664
15664
|
isUsingTts: K.usetts,
|
|
15665
15665
|
preferredLanguage: K.locale
|
|
15666
15666
|
}));
|
|
15667
|
-
this.privConversationServiceConnector.participantsListReceived && this.privConversationServiceConnector.participantsListReceived(this.privConversationServiceConnector, new a.ParticipantsListEventArgs(
|
|
15667
|
+
this.privConversationServiceConnector.participantsListReceived && this.privConversationServiceConnector.participantsListReceived(this.privConversationServiceConnector, new a.ParticipantsListEventArgs(A.roomid, A.token, A.translateTo, A.profanityFilter, A.roomProfanityFilter, A.roomLocked, A.muteAll, I, h));
|
|
15668
15668
|
break;
|
|
15669
15669
|
case "settranslatetolanguages":
|
|
15670
15670
|
this.privConversationServiceConnector.participantUpdateCommandReceived && this.privConversationServiceConnector.participantUpdateCommandReceived(this.privConversationServiceConnector, new a.ParticipantAttributeEventArgs(m.participantId, o.ConversationTranslatorCommandTypes.setTranslateToLanguages, m.value, h));
|
|
@@ -15691,14 +15691,14 @@ function Gm() {
|
|
|
15691
15691
|
this.privConversationServiceConnector.participantUpdateCommandReceived && this.privConversationServiceConnector.participantUpdateCommandReceived(this.privConversationServiceConnector, new a.ParticipantAttributeEventArgs(m.participantId, o.ConversationTranslatorCommandTypes.changeNickname, m.value, h));
|
|
15692
15692
|
break;
|
|
15693
15693
|
case "joinsession":
|
|
15694
|
-
const
|
|
15695
|
-
avatar:
|
|
15696
|
-
displayName:
|
|
15697
|
-
id:
|
|
15698
|
-
isHost:
|
|
15699
|
-
isMuted:
|
|
15700
|
-
isUsingTts:
|
|
15701
|
-
preferredLanguage:
|
|
15694
|
+
const w = n.ParticipantPayloadResponse.fromJSON(l.textBody), O = {
|
|
15695
|
+
avatar: w.avatar,
|
|
15696
|
+
displayName: w.nickname,
|
|
15697
|
+
id: w.participantId,
|
|
15698
|
+
isHost: w.ishost,
|
|
15699
|
+
isMuted: w.ismuted,
|
|
15700
|
+
isUsingTts: w.usetts,
|
|
15701
|
+
preferredLanguage: w.locale
|
|
15702
15702
|
};
|
|
15703
15703
|
this.privConversationServiceConnector.participantJoinCommandReceived && this.privConversationServiceConnector.participantJoinCommandReceived(this.privConversationServiceConnector, new a.ParticipantEventArgs(O, h));
|
|
15704
15704
|
break;
|
|
@@ -17447,28 +17447,30 @@ const _S = { class: "max-h-screen overflow-auto md:p-6 p-3 flex flex-col items-c
|
|
|
17447
17447
|
This is a role-play simulation designed to mirror real-world scenarios. Just follow the prompts, speak naturally, and let the AI guide you. Based on your responses, you’ll receive personalized feedback to help sharpen your client engagement skills.`, LS = /* @__PURE__ */ Pe({
|
|
17448
17448
|
__name: "greetingsSimulation",
|
|
17449
17449
|
setup(t) {
|
|
17450
|
-
const e = be(), r = St(), i = M(!1), c = M(!1), a = M(!0), o =
|
|
17451
|
-
|
|
17450
|
+
const e = be(), r = St(), i = M(!1), c = M(!1), a = M(!0), o = () => {
|
|
17451
|
+
e.simulationSkill === "7bd03cb6-79ad-4e2a-8a8b-88dc16147369" ? (e.beginFlag = !1, i.value = !0) : (e.beginFlag = !1, v());
|
|
17452
|
+
}, n = `Hi, get ready to test your ${e.skillNameForSimulation || "GenpactNext"} knowledge! Here’s what to expect:`, s = ` Hi ${e.learnerName || "Expert"}, and welcome!`, u = ` Hi, ${e.learnerName || "Expert"}. Get ready to test your ${e.skillNameForSimulation || "GenpactNext"} knowledge! Here’s what to expect:`, p = Lo(() => ["32b7d6d8-f7a8-40a0-ab84-8784f25897ef"].includes(e.simulationSkill) ? n : ["7bd03cb6-79ad-4e2a-8a8b-88dc16147369"].includes(e.simulationSkill) ? s : u), d = `Hi, get ready to test your ${e.skillNameForSimulation || "GenpactNext"} knowledge! Here’s what to expect:`, l = ` AI Guru will ask you a few questions on '${e.skillNameForSimulation.trim() || "GenpactNext"} '. Please provide detailed and thoughtful answers to the best of your knowledge. At the end, you'll receive a feedback, outlining your strengths and improvement areas.`, h = Lo(() => ["32b7d6d8-f7a8-40a0-ab84-8784f25897ef"].includes(e.simulationSkill) ? d : ["7bd03cb6-79ad-4e2a-8a8b-88dc16147369"].includes(e.simulationSkill) ? jS : l), v = async () => {
|
|
17453
|
+
var g, m, f, S, C, y;
|
|
17452
17454
|
console.log("firstQuestion fired==>"), a.value = !1;
|
|
17453
17455
|
try {
|
|
17454
|
-
const
|
|
17456
|
+
const P = {
|
|
17455
17457
|
question: "Hello, let's start"
|
|
17456
17458
|
};
|
|
17457
|
-
if (e.conversations.push(
|
|
17458
|
-
const
|
|
17459
|
+
if (e.conversations.push(P), e.assessmentQuestion) {
|
|
17460
|
+
const A = {
|
|
17459
17461
|
past_messages: e.past_messages,
|
|
17460
17462
|
question: "Let's begin",
|
|
17461
17463
|
bandCode: e.band_code,
|
|
17462
17464
|
userGenomeFunction: e.genomeFunctionValue,
|
|
17463
17465
|
userGenomeRole: e.genomeRoleValue,
|
|
17464
17466
|
skillId: e.simulationSkill || "937e84ef-e95d-4327-9afe-e7be2bf420eb"
|
|
17465
|
-
},
|
|
17467
|
+
}, I = M("");
|
|
17466
17468
|
e.dotLoader = !0, le.value = "listening";
|
|
17467
|
-
const
|
|
17468
|
-
e.addConversation((
|
|
17469
|
+
const w = await gt("/sqlTest", A);
|
|
17470
|
+
e.addConversation((g = w.data) == null ? void 0 : g.body, "new"), I.value = (f = (m = w == null ? void 0 : w.data) == null ? void 0 : m.body) == null ? void 0 : f.answer, e.dotLoader = !1, le.value = "bot", e.showVoiceComponent && (e.elevenLabs ? await Qd(I.value.replace(/[*#]/g, "")) : await Jd(I.value.replace(/[*#]/g, "")));
|
|
17469
17471
|
} else {
|
|
17470
17472
|
c.value = !0, e.dotLoader = !0, console.log("firstQuestion fired==> else part");
|
|
17471
|
-
const
|
|
17473
|
+
const A = {
|
|
17472
17474
|
assessmentId: e.simulationSkill,
|
|
17473
17475
|
question: "Hello, Let's begin",
|
|
17474
17476
|
past_messages: e.past_messages,
|
|
@@ -17481,28 +17483,28 @@ This is a role-play simulation designed to mirror real-world scenarios. Just fol
|
|
|
17481
17483
|
storeFinalFeedback: e.storeFinalFeedback,
|
|
17482
17484
|
storeFinalScore: e.storeFinalScore
|
|
17483
17485
|
};
|
|
17484
|
-
let
|
|
17485
|
-
Dt === "prod" ?
|
|
17486
|
-
const
|
|
17487
|
-
e.addConversation((
|
|
17486
|
+
let I;
|
|
17487
|
+
Dt === "prod" ? I = "/ManualExportActivity" : I = "/fetchCompletionDemo";
|
|
17488
|
+
const w = await gt(I, A);
|
|
17489
|
+
e.addConversation((S = w.data) == null ? void 0 : S.body, "assessment"), e.dotLoader = !1;
|
|
17488
17490
|
}
|
|
17489
|
-
} catch (
|
|
17490
|
-
if (console.error(
|
|
17491
|
-
const
|
|
17492
|
-
console.log("Error found!"), console.error((
|
|
17491
|
+
} catch (P) {
|
|
17492
|
+
if (console.error(P), Ce.isAxiosError(P)) {
|
|
17493
|
+
const A = P;
|
|
17494
|
+
console.log("Error found!"), console.error((C = A.response) == null ? void 0 : C.data), r.setError(((y = A.response) == null ? void 0 : y.data).message);
|
|
17493
17495
|
} else
|
|
17494
17496
|
r.setError("There was an issue, please contact helpmate");
|
|
17495
17497
|
}
|
|
17496
17498
|
};
|
|
17497
|
-
return (
|
|
17499
|
+
return (g, m) => (N(), j("section", null, [
|
|
17498
17500
|
ge(sp, { "is-visible": i.value }, {
|
|
17499
17501
|
default: op(() => [
|
|
17500
17502
|
R("div", _S, [
|
|
17501
17503
|
TS,
|
|
17502
17504
|
bS,
|
|
17503
17505
|
R("button", {
|
|
17504
|
-
onClick:
|
|
17505
|
-
i.value = !1,
|
|
17506
|
+
onClick: m[0] || (m[0] = (f) => {
|
|
17507
|
+
i.value = !1, v();
|
|
17506
17508
|
}),
|
|
17507
17509
|
class: "bg-[#021c42] rounded-md px-4 py-1.5 text-center text-[16px] text-white cursor-pointer italic hover:bg-blue-700"
|
|
17508
17510
|
}, " Proceed ")
|
|
@@ -17512,17 +17514,15 @@ This is a role-play simulation designed to mirror real-world scenarios. Just fol
|
|
|
17512
17514
|
}, 8, ["is-visible"]),
|
|
17513
17515
|
R("div", MS, [
|
|
17514
17516
|
R("div", OS, [
|
|
17515
|
-
R("div", DS, Ve(
|
|
17516
|
-
R("div", kS, Ve(
|
|
17517
|
+
R("div", DS, Ve(p.value), 1),
|
|
17518
|
+
R("div", kS, Ve(h.value), 1)
|
|
17517
17519
|
])
|
|
17518
17520
|
]),
|
|
17519
17521
|
k(e).beginFlag ? (N(), j("span", NS, [
|
|
17520
17522
|
R("button", {
|
|
17521
17523
|
type: "button",
|
|
17522
17524
|
id: "button",
|
|
17523
|
-
onClick:
|
|
17524
|
-
k(e).beginFlag = !1, i.value = !0;
|
|
17525
|
-
}),
|
|
17525
|
+
onClick: o,
|
|
17526
17526
|
class: "mt-4 bg-[#021c42] rounded-md px-4 py-1.5 text-center text-[16px] text-white cursor-pointer italic hover:bg-blue-700"
|
|
17527
17527
|
}, " Let's begin ")
|
|
17528
17528
|
])) : _e("", !0),
|
|
@@ -18027,7 +18027,7 @@ const _C = /* @__PURE__ */ Ft(IC, [["__scopeId", "data-v-e55a322f"]]), TC = { cl
|
|
|
18027
18027
|
console.log("ques", y), e.conversations.push(y);
|
|
18028
18028
|
const P = M("");
|
|
18029
18029
|
P.value = e.question || e.recognizedText, e.question = "";
|
|
18030
|
-
const
|
|
18030
|
+
const A = {
|
|
18031
18031
|
ohr_id: i.value,
|
|
18032
18032
|
past_messages: e.past_messages,
|
|
18033
18033
|
question: P.value || e.recognizedText,
|
|
@@ -18035,10 +18035,10 @@ const _C = /* @__PURE__ */ Ft(IC, [["__scopeId", "data-v-e55a322f"]]), TC = { cl
|
|
|
18035
18035
|
userGenomeFunction: e.genomeFunctionValue,
|
|
18036
18036
|
userGenomeRole: e.genomeRoleValue,
|
|
18037
18037
|
skillId: e.simulationSkill || "937e84ef-e95d-4327-9afe-e7be2bf420eb"
|
|
18038
|
-
}, I = await gt("/sqlTest",
|
|
18038
|
+
}, I = await gt("/sqlTest", A);
|
|
18039
18039
|
if (i.value = (h = (l = I.data) == null ? void 0 : l.body) == null ? void 0 : h.ohr, ((g = (v = I == null ? void 0 : I.data) == null ? void 0 : v.body) == null ? void 0 : g.answer) === "") {
|
|
18040
18040
|
console.log("in the if case calling api again");
|
|
18041
|
-
const O = await gt("/sqlTest",
|
|
18041
|
+
const O = await gt("/sqlTest", A);
|
|
18042
18042
|
e.addConversation((m = O.data) == null ? void 0 : m.body, "new");
|
|
18043
18043
|
} else
|
|
18044
18044
|
console.log("in the else case"), e.addConversation((f = I.data) == null ? void 0 : f.body, "new");
|
|
@@ -18121,21 +18121,21 @@ const _C = /* @__PURE__ */ Ft(IC, [["__scopeId", "data-v-e55a322f"]]), TC = { cl
|
|
|
18121
18121
|
]));
|
|
18122
18122
|
}
|
|
18123
18123
|
});
|
|
18124
|
-
const Cy = /* @__PURE__ */ Ft(Sy, [["__scopeId", "data-v-e951c0ed"]]), po = (t) => (Wn("data-v-
|
|
18125
|
-
|
|
18126
|
-
],
|
|
18124
|
+
const Cy = /* @__PURE__ */ Ft(Sy, [["__scopeId", "data-v-e951c0ed"]]), po = (t) => (Wn("data-v-bbd70acc"), t = t(), Vn(), t), yy = { class: "z-3 absolute bottom-0 mt-4 w-full bg-white px-2 md:px-6" }, Py = { class: "py-2" }, Ry = { class: "flex" }, Ey = { class: "z-1 relative w-full" }, wy = { class: "relative w-full" }, Ay = { class: "absolute inset-y-0 right-0 flex items-center md:pr-8 pr-3" }, Iy = { key: 0 }, _y = ["disabled"], Ty = /* @__PURE__ */ po(() => /* @__PURE__ */ R("path", { d: "M15.854.146a.5.5 0 0 1 .11.54l-5.819 14.547a.75.75 0 0 1-1.329.124l-3.178-4.995L.643 7.184a.75.75 0 0 1 .124-1.33L15.314.037a.5.5 0 0 1 .54.11ZM6.636 10.07l2.761 4.338L14.13 2.576 6.636 10.07Zm6.787-8.201L1.591 6.602l4.339 2.76 7.494-7.493Z" }, null, -1)), by = [
|
|
18125
|
+
Ty
|
|
18126
|
+
], My = {
|
|
18127
18127
|
key: 1,
|
|
18128
18128
|
class: "tooltip"
|
|
18129
|
-
},
|
|
18129
|
+
}, Oy = /* @__PURE__ */ po(() => /* @__PURE__ */ R("span", { class: "tooltiptext fontFormattingAMA" }, "Waiting for response", -1)), Dy = /* @__PURE__ */ po(() => /* @__PURE__ */ R("img", {
|
|
18130
|
+
class: "w-7 h-7",
|
|
18130
18131
|
src: ip,
|
|
18131
|
-
alt: "watingIcon"
|
|
18132
|
-
|
|
18133
|
-
|
|
18134
|
-
|
|
18135
|
-
|
|
18136
|
-
], ky = { class: "flex justify-start gap-2 md:gap-1" }, Ny = ["disabled"], zy = /* @__PURE__ */ po(() => /* @__PURE__ */ R("div", { class: "pt-1 text-start font-sans text-xs font-bold text-gray-500 md:text-sm" }, [
|
|
18132
|
+
alt: "watingIcon"
|
|
18133
|
+
}, null, -1)), ky = [
|
|
18134
|
+
Oy,
|
|
18135
|
+
Dy
|
|
18136
|
+
], Ny = { class: "flex justify-start gap-2 md:gap-1" }, zy = ["disabled"], jy = /* @__PURE__ */ po(() => /* @__PURE__ */ R("div", { class: "pt-1 text-start font-sans text-xs font-bold text-gray-500 md:text-sm" }, [
|
|
18137
18137
|
/* @__PURE__ */ R("div", null, " Note: AI-generated response may not be fully accurate; check with experts for critical information. ")
|
|
18138
|
-
], -1)),
|
|
18138
|
+
], -1)), Ly = /* @__PURE__ */ Pe({
|
|
18139
18139
|
__name: "NewMDMQuestionSimulation",
|
|
18140
18140
|
setup(t) {
|
|
18141
18141
|
const e = be(), r = St(), i = M("");
|
|
@@ -18193,61 +18193,62 @@ const Cy = /* @__PURE__ */ Ft(Sy, [["__scopeId", "data-v-e951c0ed"]]), po = (t)
|
|
|
18193
18193
|
R("div", Ry, [
|
|
18194
18194
|
R("div", Ey, [
|
|
18195
18195
|
R("div", wy, [
|
|
18196
|
-
|
|
18197
|
-
(N(), j("
|
|
18198
|
-
|
|
18196
|
+
R("div", Ay, [
|
|
18197
|
+
n.value ? (N(), j("div", My, ky)) : (N(), j("div", Iy, [
|
|
18198
|
+
(N(), j("svg", {
|
|
18199
|
+
disabled: k(e).beginFlag === !0,
|
|
18200
|
+
class: kt([
|
|
18201
|
+
k(e).beginFlag === !0 ? "cursor-not-allowed text-gray-600" : "cursor-pointer",
|
|
18202
|
+
"bi bi-send ml-2 h-7 w-7 rotate-45 rounded"
|
|
18203
|
+
]),
|
|
18204
|
+
onClick: d[0] || (d[0] = (l) => k(e).searchBoxdisable === !1 || k(e).beginFlag === !0 ? "" : u()),
|
|
18205
|
+
xmlns: "http://www.w3.org/2000/svg",
|
|
18206
|
+
fill: "currentColor",
|
|
18207
|
+
viewBox: "0 0 16 16"
|
|
18208
|
+
}, by, 10, _y))
|
|
18209
|
+
]))
|
|
18210
|
+
]),
|
|
18211
|
+
R("div", Ny, [
|
|
18212
|
+
Jo(R("input", {
|
|
18213
|
+
onKeyup: [
|
|
18214
|
+
d[1] || (d[1] = cp((l) => n.value ? "" : u(), ["enter"])),
|
|
18215
|
+
d[3] || (d[3] = (l) => s(k(e).question))
|
|
18216
|
+
],
|
|
18217
|
+
disabled: k(e).beginFlag === !0 || k(e).searchBoxdisable === !1,
|
|
18199
18218
|
class: kt([
|
|
18200
|
-
k(e).beginFlag === !0 ? "cursor-not-allowed
|
|
18201
|
-
"
|
|
18219
|
+
k(e).beginFlag === !0 || k(e).searchBoxdisable === !1 ? "cursor-not-allowed bg-[#e5e7eb]" : "bg-white",
|
|
18220
|
+
"fontFormattingAMA rounded-lg flex w-full border border-genpact-darkest-teal py-[18px] pl-[30px] pr-[70px]"
|
|
18202
18221
|
]),
|
|
18203
|
-
|
|
18204
|
-
|
|
18205
|
-
|
|
18206
|
-
|
|
18207
|
-
|
|
18208
|
-
|
|
18209
|
-
|
|
18210
|
-
|
|
18211
|
-
Jo(R("textarea", {
|
|
18212
|
-
rows: "2",
|
|
18213
|
-
onKeyup: [
|
|
18214
|
-
d[1] || (d[1] = cp((l) => u(), ["enter"])),
|
|
18215
|
-
d[3] || (d[3] = (l) => s(k(e).question))
|
|
18216
|
-
],
|
|
18217
|
-
disabled: k(e).beginFlag === !0,
|
|
18218
|
-
class: kt([
|
|
18219
|
-
k(e).beginFlag === !0 || k(e).searchBoxdisable === !1 ? "cursor-not-allowed bg-[#e5e7eb]" : "",
|
|
18220
|
-
"fontFormattingAMA flex w-full border border-genpact-darkest-teal py-2 pl-2 pr-16 focus:outline-none sm:text-[15px] md:py-3 md:pl-6 lg:py-3 lg:pl-6"
|
|
18221
|
-
]),
|
|
18222
|
-
"onUpdate:modelValue": d[2] || (d[2] = (l) => k(e).question = l),
|
|
18223
|
-
placeholder: "Write your answer here...",
|
|
18224
|
-
type: "text",
|
|
18225
|
-
ref_key: "AiSearch",
|
|
18226
|
-
ref: c
|
|
18227
|
-
}, null, 42, Ny), [
|
|
18228
|
-
[ap, k(e).question]
|
|
18222
|
+
"onUpdate:modelValue": d[2] || (d[2] = (l) => k(e).question = l),
|
|
18223
|
+
placeholder: "Write your answer here...",
|
|
18224
|
+
type: "text",
|
|
18225
|
+
ref_key: "AiSearch",
|
|
18226
|
+
ref: c
|
|
18227
|
+
}, null, 42, zy), [
|
|
18228
|
+
[ap, k(e).question]
|
|
18229
|
+
])
|
|
18229
18230
|
])
|
|
18230
18231
|
]),
|
|
18231
|
-
|
|
18232
|
+
jy
|
|
18232
18233
|
])
|
|
18233
18234
|
])
|
|
18234
18235
|
])
|
|
18235
18236
|
]));
|
|
18236
18237
|
}
|
|
18237
18238
|
});
|
|
18238
|
-
const
|
|
18239
|
+
const xy = /* @__PURE__ */ Ft(Ly, [["__scopeId", "data-v-bbd70acc"]]), Xd = (t) => (Wn("data-v-904e6953"), t = t(), Vn(), t), qy = /* @__PURE__ */ Xd(() => /* @__PURE__ */ R("div", { class: "max-h-screen overflow-auto md:p-6 p-3 flex flex-col items-center justify-center gap-4 text-[#161916]" }, [
|
|
18239
18240
|
/* @__PURE__ */ R("div", { class: "text-2xl font-[700]" }, "Assessment Access Required"),
|
|
18240
18241
|
/* @__PURE__ */ R("h3", { className: "text-lg font-medium my-6 text-center text-gray-500" }, " You don't have access to this assessment. Please contact to administrator ")
|
|
18241
|
-
], -1)),
|
|
18242
|
+
], -1)), By = {
|
|
18242
18243
|
key: 1,
|
|
18243
18244
|
className: "h-screen flex flex-col items-stretch"
|
|
18244
|
-
},
|
|
18245
|
+
}, Fy = { class: "z-20" }, Uy = /* @__PURE__ */ Xd(() => /* @__PURE__ */ R("div", { class: "mt-6" }, null, -1)), Hy = { class: "flex h-full items-stretch overflow-hidden" }, Wy = { class: "mx-auto w-11/12 bg-white" }, Vy = { class: "relative flex h-full w-full flex-col items-stretch overflow-hidden" }, $y = {
|
|
18245
18246
|
key: 0,
|
|
18246
18247
|
class: "relative"
|
|
18247
|
-
},
|
|
18248
|
+
}, Ky = {
|
|
18248
18249
|
key: 1,
|
|
18249
18250
|
class: "relative"
|
|
18250
|
-
},
|
|
18251
|
+
}, Jy = /* @__PURE__ */ Pe({
|
|
18251
18252
|
__name: "AssessmentSimulation",
|
|
18252
18253
|
setup(t) {
|
|
18253
18254
|
const e = M(!0), r = M(!1), i = M(!1);
|
|
@@ -18261,12 +18262,12 @@ const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t)
|
|
|
18261
18262
|
console.log("uuid. value", g), e.value = !0, await S(), await C(), c.assessmentQuestion || (console.log("getAssessmentDetails========================>"), await f()), e.value = !1;
|
|
18262
18263
|
});
|
|
18263
18264
|
const f = async () => {
|
|
18264
|
-
var y, P,
|
|
18265
|
+
var y, P, A, I, w, O, F, Q, G, K, $t, ii, si, oi, ai, ci, ui, pi, di, li, hi, vi, fi, gi, mi, Si, Ci, yi, Pi, Ri, Ei, wi, Ai, Ii, _i, Ti, bi, Kt, Jt, Mi, Oi, Di, ki, Ni, zi, ji, Li, Pt;
|
|
18265
18266
|
console.log("Start getAssessmentDetails");
|
|
18266
18267
|
try {
|
|
18267
18268
|
Dt === "prod" ? p = `https://api-v2.genpact.com/api/my-profile/assessment-master/getAssessmentDetailsById?id=${c.simulationSkill}` : p = `https://api-v2-dev.genpact.com/api/my-profile/assessment-master/getAssessmentDetailsById?id=${c.simulationSkill}`;
|
|
18268
18269
|
const _ = await xn(p);
|
|
18269
|
-
o.value = (P = (y = _ == null ? void 0 : _.data) == null ? void 0 : y.data) == null ? void 0 : P.name, c.skillNameForSimulation = o.value, c.generatedPrompt = (I = (
|
|
18270
|
+
o.value = (P = (y = _ == null ? void 0 : _.data) == null ? void 0 : y.data) == null ? void 0 : P.name, c.skillNameForSimulation = o.value, c.generatedPrompt = (I = (A = _ == null ? void 0 : _.data) == null ? void 0 : A.data) == null ? void 0 : I.generatedPrompt, c.assessmentType = (O = (w = _ == null ? void 0 : _.data) == null ? void 0 : w.data) == null ? void 0 : O.assessmentType, s.value = (Q = (F = _ == null ? void 0 : _.data) == null ? void 0 : F.data) == null ? void 0 : Q.status, h.value = (K = (G = _ == null ? void 0 : _.data) == null ? void 0 : G.data) == null ? void 0 : K.editMode, c.empSeniority = (ii = ($t = _ == null ? void 0 : _.data) == null ? void 0 : $t.data) == null ? void 0 : ii.empSeniority, c.functionRole = (oi = (si = _ == null ? void 0 : _.data) == null ? void 0 : si.data) == null ? void 0 : oi.functionRole, d = (ci = (ai = _ == null ? void 0 : _.data) == null ? void 0 : ai.data) == null ? void 0 : ci.accessBand, l = (pi = (ui = _ == null ? void 0 : _.data) == null ? void 0 : ui.data) == null ? void 0 : pi.accessOhrList;
|
|
18270
18271
|
const Le = (li = (di = _ == null ? void 0 : _.data) == null ? void 0 : di.data) == null ? void 0 : li.createdBy;
|
|
18271
18272
|
if (c.mcqType = (vi = (hi = _ == null ? void 0 : _.data) == null ? void 0 : hi.data) == null ? void 0 : vi.mcqTypes, c.storeFinalFeedback = (gi = (fi = _ == null ? void 0 : _.data) == null ? void 0 : fi.data) == null ? void 0 : gi.storeFinalFeedback, c.storeFinalScore = (Si = (mi = _ == null ? void 0 : _.data) == null ? void 0 : mi.data) == null ? void 0 : Si.storeFinalScore, s.value === "inactive")
|
|
18272
18273
|
r.value = !0;
|
|
@@ -18318,7 +18319,7 @@ const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t)
|
|
|
18318
18319
|
a.setError("There was an issue, please contact helpmate");
|
|
18319
18320
|
}
|
|
18320
18321
|
}, S = async () => {
|
|
18321
|
-
var y, P,
|
|
18322
|
+
var y, P, A, I, w;
|
|
18322
18323
|
try {
|
|
18323
18324
|
if (g === null)
|
|
18324
18325
|
throw {
|
|
@@ -18339,16 +18340,16 @@ const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t)
|
|
|
18339
18340
|
} catch (O) {
|
|
18340
18341
|
if (console.error(O), e.value = !1, Ce.isAxiosError(O) && ((y = O.response) != null && y.status) || O && typeof O == "object" && "response" in O && O.response && typeof O.response == "object" && "status" in O.response && O.response.status) {
|
|
18341
18342
|
const F = O;
|
|
18342
|
-
if (console.log("Error found!"), console.error((P = F.response) == null ? void 0 : P.data), ((
|
|
18343
|
+
if (console.log("Error found!"), console.error((P = F.response) == null ? void 0 : P.data), ((A = F.response) == null ? void 0 : A.status) === 424 || ((I = F.response) == null ? void 0 : I.status) === 425) {
|
|
18343
18344
|
i.value = !0;
|
|
18344
18345
|
return;
|
|
18345
18346
|
}
|
|
18346
|
-
a.setError(((
|
|
18347
|
+
a.setError(((w = F.response) == null ? void 0 : w.data).message);
|
|
18347
18348
|
} else
|
|
18348
18349
|
a.setError("There was an issue, please contact helpmate");
|
|
18349
18350
|
}
|
|
18350
18351
|
}, C = async () => {
|
|
18351
|
-
var y, P,
|
|
18352
|
+
var y, P, A, I, w, O, F, Q;
|
|
18352
18353
|
try {
|
|
18353
18354
|
const G = await Ce({
|
|
18354
18355
|
url: "https://api-v2-dev.genpact.com/api/delphi/v2/myprofile",
|
|
@@ -18358,7 +18359,7 @@ const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t)
|
|
|
18358
18359
|
},
|
|
18359
18360
|
method: "get"
|
|
18360
18361
|
});
|
|
18361
|
-
c.genomeFunctionValue = (
|
|
18362
|
+
c.genomeFunctionValue = (A = (P = (y = G == null ? void 0 : G.data) == null ? void 0 : y.role_function) == null ? void 0 : P.function) == null ? void 0 : A.name, c.genomeRoleValue = (O = (w = (I = G == null ? void 0 : G.data) == null ? void 0 : I.role_function) == null ? void 0 : w.role) == null ? void 0 : O.name, console.log("profile data", c.genomeFunctionValue, c.genomeRoleValue);
|
|
18362
18363
|
} catch (G) {
|
|
18363
18364
|
if (console.error(G), e.value = !1, Ce.isAxiosError(G)) {
|
|
18364
18365
|
const K = G;
|
|
@@ -18370,28 +18371,28 @@ const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t)
|
|
|
18370
18371
|
return (y, P) => (N(), j("section", null, [
|
|
18371
18372
|
ge(sp, { "is-visible": i.value }, {
|
|
18372
18373
|
default: op(() => [
|
|
18373
|
-
|
|
18374
|
+
qy
|
|
18374
18375
|
]),
|
|
18375
18376
|
_: 1
|
|
18376
18377
|
}, 8, ["is-visible"]),
|
|
18377
18378
|
R("div", null, [
|
|
18378
|
-
e.value ? (N(), Al(yl, { key: 0 })) : (N(), j("main",
|
|
18379
|
-
R("div",
|
|
18379
|
+
e.value ? (N(), Al(yl, { key: 0 })) : (N(), j("main", By, [
|
|
18380
|
+
R("div", Fy, [
|
|
18380
18381
|
ge(zl)
|
|
18381
18382
|
]),
|
|
18382
|
-
|
|
18383
|
-
R("div",
|
|
18384
|
-
R("div",
|
|
18385
|
-
R("div",
|
|
18383
|
+
Uy,
|
|
18384
|
+
R("div", Hy, [
|
|
18385
|
+
R("div", Wy, [
|
|
18386
|
+
R("div", Vy, [
|
|
18386
18387
|
ge(_C),
|
|
18387
18388
|
R("div", {
|
|
18388
18389
|
class: kt(k(c).showVoiceComponent === !0 ? "mt-[160px]" : "mt-[120px]")
|
|
18389
18390
|
}, null, 2)
|
|
18390
18391
|
]),
|
|
18391
|
-
k(c).assessmentQuestion ? (N(), j("div",
|
|
18392
|
+
k(c).assessmentQuestion ? (N(), j("div", $y, [
|
|
18392
18393
|
ge(Cy)
|
|
18393
|
-
])) : (N(), j("div",
|
|
18394
|
-
ge(
|
|
18394
|
+
])) : (N(), j("div", Ky, [
|
|
18395
|
+
ge(xy)
|
|
18395
18396
|
]))
|
|
18396
18397
|
])
|
|
18397
18398
|
])
|
|
@@ -18400,7 +18401,7 @@ const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t)
|
|
|
18400
18401
|
]));
|
|
18401
18402
|
}
|
|
18402
18403
|
});
|
|
18403
|
-
const
|
|
18404
|
+
const RR = /* @__PURE__ */ Ft(Jy, [["__scopeId", "data-v-904e6953"]]);
|
|
18404
18405
|
export {
|
|
18405
|
-
|
|
18406
|
+
RR as default
|
|
18406
18407
|
};
|