@genpact/genome.mfe.mgtwin-app 1.1.267-alpha → 1.1.269-alpha

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,8 +1,8 @@
1
- import { _ as Cl, a as ip, L as yl } from "./LoadingSimulation-e1c451ea.mjs";
2
- import { _ as Pl, a as sp, Q as Ot, M as Aa } from "./Modal-abe82066.mjs";
3
- import { d as Pe, r as M, H as Dt, o as N, c as j, a as R, k as St, u as be, C as Bt, q as Ce, y as Lo, h as ge, s as op, t as Ve, b as k, f as _e, i as zn, I as Vi, F as xo, e as qo, w as Jo, J as Rl, p as Wn, g as Vn, G as Gi, z as El, K as wl, n as kt, m as ap, L as cp, l as Al } from "./main-b1d63f30.mjs";
4
- import { g as xn, p as gt } from "./api-91aec5f8.mjs";
5
- import { c as B, a as Il, r as _l } from "./_commonjsHelpers-89b7dddd.mjs";
1
+ import { _ as Cl, a as ip, L as yl } from "./LoadingSimulation-27c136fb.mjs";
2
+ import { _ as Pl, a as sp, Q as Ot, M as Aa } from "./Modal-cdd2991e.mjs";
3
+ import { d as Pe, r as M, H as Dt, o as N, c as j, a as R, k as St, u as be, C as Bt, q as Ce, y as Lo, h as ge, s as op, t as Ve, b as k, f as _e, i as zn, I as Vi, F as xo, e as qo, w as Jo, J as Rl, p as Wn, g as Vn, G as Gi, z as El, K as wl, n as kt, m as ap, L as cp, l as Al } from "./main-fb55944a.mjs";
4
+ import { g as xn, p as gt } from "./api-0bcaa8e3.mjs";
5
+ import { c as B, a as Il, r as _l } from "./_commonjsHelpers-e1f9766e.mjs";
6
6
  import { _ as Ft } from "./_plugin-vue_export-helper-dad06003.mjs";
7
7
  import "./loadernew-ac153a51.mjs";
8
8
  const Tl = { class: "flex h-14 w-full flex-1 items-center justify-between bg-[#021c42] md:pr-6 lg:pr-6" }, bl = ["href"], Ml = /* @__PURE__ */ R("img", {
@@ -6514,12 +6514,12 @@ function Kf() {
6514
6514
  p.token != null && p.token !== "" && (y[p.headerName] = p.token);
6515
6515
  let P = u.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Endpoint, "");
6516
6516
  if (!P) {
6517
- const I = i.ConnectionFactoryBase.getHostSuffix(v), A = u.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Host, `wss://${v}.${n.BaseUrl}${I}`);
6518
- P = `${A.endsWith("/") ? A : A + "/"}${S}${n.ApiKey}/${C}`;
6517
+ const I = i.ConnectionFactoryBase.getHostSuffix(v), w = u.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Host, `wss://${v}.${n.BaseUrl}${I}`);
6518
+ P = `${w.endsWith("/") ? w : w + "/"}${S}${n.ApiKey}/${C}`;
6519
6519
  }
6520
6520
  this.setCommonUrlParams(u, f, P);
6521
- const w = u.parameters.getProperty("SPEECH-EnableWebsocketCompression", "false") === "true";
6522
- return Promise.resolve(new t.WebsocketConnection(P, f, y, new c.WebsocketMessageFormatter(), t.ProxyInfo.fromRecognizerConfig(u), w, d));
6521
+ const A = u.parameters.getProperty("SPEECH-EnableWebsocketCompression", "false") === "true";
6522
+ return Promise.resolve(new t.WebsocketConnection(P, f, y, new c.WebsocketMessageFormatter(), t.ProxyInfo.fromRecognizerConfig(u), A, d));
6523
6523
  }
6524
6524
  }
6525
6525
  return Tr.DialogConnectionFactory = n, n.ApiKey = "api", n.BaseUrl = "convai.speech", Tr;
@@ -8393,10 +8393,10 @@ function Sg() {
8393
8393
  if (h) {
8394
8394
  f = p.parameters.getProperty(i.PropertyId.SpeechServiceConnection_Endpoint), f || (f = "wss://" + p.parameters.getProperty(i.PropertyId.SpeechServiceConnection_Host, "transcribe.{region}.cts.speech{hostSuffix}") + "{path}"), f = e.StringUtils.formatString(f, m);
8395
8395
  const P = new URL(f);
8396
- P.searchParams.forEach((I, A) => {
8397
- S[A] = I;
8398
- }), new n.TranscriberConnectionFactory().setQueryParams(S, p, f), S[a.QueryParameterNames.CtsMeetingId] = m[a.QueryParameterNames.CtsMeetingId], S[a.QueryParameterNames.CtsDeviceId] = m[a.QueryParameterNames.CtsDeviceId], v.isHost || (S[a.QueryParameterNames.CtsIsParticipant] = ""), a.QueryParameterNames.Format in S || (S[a.QueryParameterNames.Format] = "simple"), P.searchParams.forEach((I, A) => {
8399
- P.searchParams.set(A, S[A]), delete S[A];
8396
+ P.searchParams.forEach((I, w) => {
8397
+ S[w] = I;
8398
+ }), new n.TranscriberConnectionFactory().setQueryParams(S, p, f), S[a.QueryParameterNames.CtsMeetingId] = m[a.QueryParameterNames.CtsMeetingId], S[a.QueryParameterNames.CtsDeviceId] = m[a.QueryParameterNames.CtsDeviceId], v.isHost || (S[a.QueryParameterNames.CtsIsParticipant] = ""), a.QueryParameterNames.Format in S || (S[a.QueryParameterNames.Format] = "simple"), P.searchParams.forEach((I, w) => {
8399
+ P.searchParams.set(w, S[w]), delete S[w];
8400
8400
  }), f = P.toString();
8401
8401
  } else {
8402
8402
  const P = new n.TranslationConnectionFactory();
@@ -8518,8 +8518,8 @@ function Cg() {
8518
8518
  for (const y of S) {
8519
8519
  const P = this.privProperties.getProperty(y);
8520
8520
  if (P) {
8521
- const w = typeof y == "string" ? y : a.PropertyId[y];
8522
- this.privSpeechTranslationConfig.setProperty(w, P);
8521
+ const A = typeof y == "string" ? y : a.PropertyId[y];
8522
+ this.privSpeechTranslationConfig.setProperty(A, P);
8523
8523
  }
8524
8524
  }
8525
8525
  const C = JSON.parse(this.privProperties.getProperty(e.ServicePropertiesPropertyName, "{}"));
@@ -11425,19 +11425,19 @@ function E() {
11425
11425
  Object.defineProperty(t, "TranslationRecognitionResult", { enumerable: !0, get: function() {
11426
11426
  return P.TranslationRecognitionResult;
11427
11427
  } });
11428
- var w = _s;
11428
+ var A = _s;
11429
11429
  Object.defineProperty(t, "TranslationSynthesisResult", { enumerable: !0, get: function() {
11430
- return w.TranslationSynthesisResult;
11430
+ return A.TranslationSynthesisResult;
11431
11431
  } });
11432
11432
  var I = Vp;
11433
11433
  Object.defineProperty(t, "ResultReason", { enumerable: !0, get: function() {
11434
11434
  return I.ResultReason;
11435
11435
  } });
11436
- var A = Of();
11436
+ var w = Of();
11437
11437
  Object.defineProperty(t, "SpeechConfig", { enumerable: !0, get: function() {
11438
- return A.SpeechConfig;
11438
+ return w.SpeechConfig;
11439
11439
  } }), Object.defineProperty(t, "SpeechConfigImpl", { enumerable: !0, get: function() {
11440
- return A.SpeechConfigImpl;
11440
+ return w.SpeechConfigImpl;
11441
11441
  } });
11442
11442
  var O = Df();
11443
11443
  Object.defineProperty(t, "SpeechTranslationConfig", { enumerable: !0, get: function() {
@@ -12330,8 +12330,8 @@ function tm() {
12330
12330
  throw new e.ArgumentNullError("recognizerConfig");
12331
12331
  this.privEnableSpeakerId = y.isSpeakerDiarizationEnabled, this.privMustReportEndOfStream = !1, this.privAuthentication = f, this.privConnectionFactory = S, this.privAudioSource = C, this.privRecognizerConfig = y, this.privIsDisposed = !1, this.privRecognizer = P, this.privRequestSession = new i.RequestSession(this.privAudioSource.id()), this.privConnectionEvents = new e.EventSource(), this.privServiceEvents = new e.EventSource(), this.privDynamicGrammar = new i.DynamicGrammarBuilder(), this.privSpeechContext = new i.SpeechContext(this.privDynamicGrammar), this.privAgentConfig = new i.AgentConfig(), this.privRecognizerConfig.parameters.getProperty(r.PropertyId.WebWorkerLoadType, "on").toLowerCase() === "on" && typeof Blob < "u" && typeof Worker < "u" ? this.privSetTimeout = e.Timeout.setTimeout : (typeof window < "u" && (this.privSetTimeout = window.setTimeout.bind(window)), typeof globalThis < "u" && (this.privSetTimeout = globalThis.setTimeout.bind(globalThis))), this.connectionEvents.attach((I) => {
12332
12332
  if (I.name === "ConnectionClosedEvent") {
12333
- const A = I;
12334
- (A.statusCode === 1003 || A.statusCode === 1007 || A.statusCode === 1002 || A.statusCode === 4e3 || this.privRequestSession.numConnectionAttempts > this.privRecognizerConfig.maxRetryCount) && this.cancelRecognitionLocal(r.CancellationReason.Error, A.statusCode === 1007 ? r.CancellationErrorCode.BadRequestParameters : r.CancellationErrorCode.ConnectionFailure, `${A.reason} websocket error code: ${A.statusCode}`);
12333
+ const w = I;
12334
+ (w.statusCode === 1003 || w.statusCode === 1007 || w.statusCode === 1002 || w.statusCode === 4e3 || this.privRequestSession.numConnectionAttempts > this.privRecognizerConfig.maxRetryCount) && this.cancelRecognitionLocal(r.CancellationReason.Error, w.statusCode === 1007 ? r.CancellationErrorCode.BadRequestParameters : r.CancellationErrorCode.ConnectionFailure, `${w.reason} websocket error code: ${w.statusCode}`);
12335
12335
  }
12336
12336
  }), this.privEnableSpeakerId && (this.privDiarizationSessionId = (0, e.createNoDashGuid)());
12337
12337
  }
@@ -12349,14 +12349,14 @@ function tm() {
12349
12349
  targetLanguages: S
12350
12350
  }, y !== void 0 && (this.privSpeechContext.getContext().translation.category = y), C !== void 0) {
12351
12351
  const I = {};
12352
- for (const A of S)
12353
- I[A] = C;
12352
+ for (const w of S)
12353
+ I[w] = C;
12354
12354
  this.privSpeechContext.getContext().synthesis = {
12355
12355
  defaultVoices: I
12356
12356
  };
12357
12357
  }
12358
- const w = this.privSpeechContext.getContext().phraseDetection || {};
12359
- w.onSuccess = { action: v.NextAction.Translate }, w.onInterim = { action: v.NextAction.Translate }, this.privSpeechContext.getContext().phraseDetection = w;
12358
+ const A = this.privSpeechContext.getContext().phraseDetection || {};
12359
+ A.onSuccess = { action: v.NextAction.Translate }, A.onInterim = { action: v.NextAction.Translate }, this.privSpeechContext.getContext().phraseDetection = A;
12360
12360
  }
12361
12361
  }
12362
12362
  setSpeechSegmentationTimeoutJson() {
@@ -12368,43 +12368,43 @@ function tm() {
12368
12368
  let P = !1;
12369
12369
  if (C !== void 0) {
12370
12370
  P = !0;
12371
- let w = a.SegmentationMode.Normal;
12371
+ let A = a.SegmentationMode.Normal;
12372
12372
  switch (C.toLowerCase()) {
12373
12373
  case "default":
12374
12374
  break;
12375
12375
  case "time":
12376
- w = a.SegmentationMode.Custom;
12376
+ A = a.SegmentationMode.Custom;
12377
12377
  break;
12378
12378
  case "semantic":
12379
- w = a.SegmentationMode.Semantic;
12379
+ A = a.SegmentationMode.Semantic;
12380
12380
  break;
12381
12381
  }
12382
- y.mode = w;
12382
+ y.mode = A;
12383
12383
  }
12384
12384
  if (f !== void 0) {
12385
12385
  P = !0;
12386
- const w = parseInt(f, 10);
12387
- y.mode = a.SegmentationMode.Custom, y.segmentationSilenceTimeoutMs = w;
12386
+ const A = parseInt(f, 10);
12387
+ y.mode = a.SegmentationMode.Custom, y.segmentationSilenceTimeoutMs = A;
12388
12388
  }
12389
12389
  if (S !== void 0) {
12390
12390
  P = !0;
12391
- const w = parseInt(S, 10);
12392
- y.mode = a.SegmentationMode.Custom, y.segmentationForcedTimeoutMs = w;
12391
+ const A = parseInt(S, 10);
12392
+ y.mode = a.SegmentationMode.Custom, y.segmentationForcedTimeoutMs = A;
12393
12393
  }
12394
12394
  if (P) {
12395
- const w = this.privSpeechContext.getContext().phraseDetection || {};
12396
- switch (w.mode = this.recognitionMode, this.recognitionMode) {
12395
+ const A = this.privSpeechContext.getContext().phraseDetection || {};
12396
+ switch (A.mode = this.recognitionMode, this.recognitionMode) {
12397
12397
  case o.RecognitionMode.Conversation:
12398
- w.conversation = w.conversation ?? { segmentation: {} }, w.conversation.segmentation = y;
12398
+ A.conversation = A.conversation ?? { segmentation: {} }, A.conversation.segmentation = y;
12399
12399
  break;
12400
12400
  case o.RecognitionMode.Interactive:
12401
- w.interactive = w.interactive ?? { segmentation: {} }, w.interactive.segmentation = y;
12401
+ A.interactive = A.interactive ?? { segmentation: {} }, A.interactive.segmentation = y;
12402
12402
  break;
12403
12403
  case o.RecognitionMode.Dictation:
12404
- w.dictation = w.dictation ?? {}, w.dictation.segmentation = y;
12404
+ A.dictation = A.dictation ?? {}, A.dictation.segmentation = y;
12405
12405
  break;
12406
12406
  }
12407
- this.privSpeechContext.getContext().phraseDetection = w;
12407
+ this.privSpeechContext.getContext().phraseDetection = A;
12408
12408
  }
12409
12409
  }
12410
12410
  setLanguageIdJson() {
@@ -12485,27 +12485,27 @@ function tm() {
12485
12485
  return;
12486
12486
  }
12487
12487
  if (this.privConnectionConfigurationPromise = void 0, this.privRecognizerConfig.recognitionMode = f, this.privRecognizerConfig.recognitionEndpointVersion === "2") {
12488
- const A = this.privSpeechContext.getContext().phraseDetection || {};
12489
- A.mode = f, this.privSpeechContext.getContext().phraseDetection = A;
12488
+ const w = this.privSpeechContext.getContext().phraseDetection || {};
12489
+ w.mode = f, this.privSpeechContext.getContext().phraseDetection = w;
12490
12490
  }
12491
12491
  this.setLanguageIdJson(), this.setTranslationJson(), this.privRecognizerConfig.autoDetectSourceLanguages !== void 0 && this.privRecognizerConfig.parameters.getProperty(r.PropertyId.SpeechServiceConnection_TranslationToLanguages, void 0) !== void 0 && this.setupTranslationWithLanguageId(), this.setSpeechSegmentationTimeoutJson(), this.setOutputDetailLevelJson(), this.privSuccessCallback = S, this.privErrorCallback = C, this.privRequestSession.startNewRecognition(), this.privRequestSession.listenForServiceTelemetry(this.privAudioSource.events);
12492
12492
  const y = this.connectImpl();
12493
12493
  let P;
12494
12494
  try {
12495
- const A = await this.audioSource.attach(this.privRequestSession.audioNodeId), O = await this.audioSource.format, F = await this.audioSource.deviceInfo;
12496
- this.privIsLiveAudio = F.type && F.type === i.type.Microphones, P = new t.ReplayableAudioNode(A, O.avgBytesPerSec), await this.privRequestSession.onAudioSourceAttachCompleted(P, !1), this.privRecognizerConfig.SpeechServiceConfig.Context.audio = { source: F };
12497
- } catch (A) {
12498
- throw await this.privRequestSession.onStopRecognizing(), A;
12495
+ const w = await this.audioSource.attach(this.privRequestSession.audioNodeId), O = await this.audioSource.format, F = await this.audioSource.deviceInfo;
12496
+ this.privIsLiveAudio = F.type && F.type === i.type.Microphones, P = new t.ReplayableAudioNode(w, O.avgBytesPerSec), await this.privRequestSession.onAudioSourceAttachCompleted(P, !1), this.privRecognizerConfig.SpeechServiceConfig.Context.audio = { source: F };
12497
+ } catch (w) {
12498
+ throw await this.privRequestSession.onStopRecognizing(), w;
12499
12499
  }
12500
12500
  try {
12501
12501
  await y;
12502
- } catch (A) {
12503
- await this.cancelRecognitionLocal(r.CancellationReason.Error, r.CancellationErrorCode.ConnectionFailure, A);
12502
+ } catch (w) {
12503
+ await this.cancelRecognitionLocal(r.CancellationReason.Error, r.CancellationErrorCode.ConnectionFailure, w);
12504
12504
  return;
12505
12505
  }
12506
- const w = new r.SessionEventArgs(this.privRequestSession.sessionId);
12507
- this.privRecognizer.sessionStarted && this.privRecognizer.sessionStarted(this.privRecognizer, w), this.receiveMessage(), this.sendAudio(P).catch(async (A) => {
12508
- await this.cancelRecognitionLocal(r.CancellationReason.Error, r.CancellationErrorCode.RuntimeError, A);
12506
+ const A = new r.SessionEventArgs(this.privRequestSession.sessionId);
12507
+ this.privRecognizer.sessionStarted && this.privRecognizer.sessionStarted(this.privRecognizer, A), this.receiveMessage(), this.sendAudio(P).catch(async (w) => {
12508
+ await this.cancelRecognitionLocal(r.CancellationReason.Error, r.CancellationErrorCode.RuntimeError, w);
12509
12509
  });
12510
12510
  }
12511
12511
  async stopRecognizing() {
@@ -12594,10 +12594,10 @@ function tm() {
12594
12594
  this.privRecognizer.speechStartDetected && this.privRecognizer.speechStartDetected(this.privRecognizer, P);
12595
12595
  break;
12596
12596
  case "speech.enddetected":
12597
- let w;
12598
- C.textBody.length > 0 ? w = C.textBody : w = "{ Offset: 0 }";
12599
- const I = i.SpeechDetected.fromJSON(w, this.privRequestSession.currentTurnAudioOffset), A = new r.RecognitionEventArgs(I.Offset + this.privRequestSession.currentTurnAudioOffset, this.privRequestSession.sessionId);
12600
- this.privRecognizer.speechEndDetected && this.privRecognizer.speechEndDetected(this.privRecognizer, A);
12597
+ let A;
12598
+ C.textBody.length > 0 ? A = C.textBody : A = "{ Offset: 0 }";
12599
+ const I = i.SpeechDetected.fromJSON(A, this.privRequestSession.currentTurnAudioOffset), w = new r.RecognitionEventArgs(I.Offset + this.privRequestSession.currentTurnAudioOffset, this.privRequestSession.sessionId);
12600
+ this.privRecognizer.speechEndDetected && this.privRecognizer.speechEndDetected(this.privRecognizer, w);
12601
12601
  break;
12602
12602
  case "turn.end":
12603
12603
  await this.sendTelemetryData(), this.privRequestSession.isSpeechEnded && this.privMustReportEndOfStream && (this.privMustReportEndOfStream = !1, await this.cancelRecognitionLocal(r.CancellationReason.EndOfStream, r.CancellationErrorCode.NoError, void 0));
@@ -12641,8 +12641,8 @@ function tm() {
12641
12641
  if (C) {
12642
12642
  const y = this.privRecognizerConfig.sourceLanguageModels;
12643
12643
  if (y !== void 0 && y.length > 0) {
12644
- const w = this.privSpeechContext.getContext().phraseDetection || {};
12645
- w.customModels = y, this.privSpeechContext.getContext().phraseDetection = w;
12644
+ const A = this.privSpeechContext.getContext().phraseDetection || {};
12645
+ A.customModels = y, this.privSpeechContext.getContext().phraseDetection = A;
12646
12646
  }
12647
12647
  this.privRecognizerConfig.parameters.getProperty(r.PropertyId.SpeechServiceConnection_TranslationVoice, void 0) !== void 0 && (C.onSuccess = { action: n.NextAction.Synthesize }, C.onPassthrough = { action: n.NextAction.Synthesize });
12648
12648
  }
@@ -12688,14 +12688,14 @@ function tm() {
12688
12688
  const S = await this.audioSource.format;
12689
12689
  this.privAverageBytesPerMs = S.avgBytesPerSec / 1e3;
12690
12690
  let C = Date.now();
12691
- const y = this.privRecognizerConfig.parameters.getProperty("SPEECH-TransmitLengthBeforThrottleMs", "5000"), P = S.avgBytesPerSec / 1e3 * parseInt(y, 10), w = this.privRequestSession.recogNumber, I = async () => {
12692
- if (!this.privIsDisposed && !this.privRequestSession.isSpeechEnded && this.privRequestSession.isRecognizing && this.privRequestSession.recogNumber === w) {
12693
- const A = await this.fetchConnection(), O = await f.read();
12691
+ const y = this.privRecognizerConfig.parameters.getProperty("SPEECH-TransmitLengthBeforThrottleMs", "5000"), P = S.avgBytesPerSec / 1e3 * parseInt(y, 10), A = this.privRequestSession.recogNumber, I = async () => {
12692
+ if (!this.privIsDisposed && !this.privRequestSession.isSpeechEnded && this.privRequestSession.isRecognizing && this.privRequestSession.recogNumber === A) {
12693
+ const w = await this.fetchConnection(), O = await f.read();
12694
12694
  if (this.privRequestSession.isSpeechEnded)
12695
12695
  return;
12696
12696
  let F, Q;
12697
- if (!O || O.isEnd ? (F = null, Q = 0) : (F = O.buffer, this.privRequestSession.onAudioSent(F.byteLength), P >= this.privRequestSession.bytesSent ? Q = 0 : Q = Math.max(0, C - Date.now())), Q !== 0 && await this.delay(Q), F !== null && (C = Date.now() + F.byteLength * 1e3 / (S.avgBytesPerSec * 2)), !this.privIsDisposed && !this.privRequestSession.isSpeechEnded && this.privRequestSession.isRecognizing && this.privRequestSession.recogNumber === w)
12698
- if (A.send(new c.SpeechConnectionMessage(e.MessageType.Binary, "audio", this.privRequestSession.requestId, null, F)).catch(() => {
12697
+ if (!O || O.isEnd ? (F = null, Q = 0) : (F = O.buffer, this.privRequestSession.onAudioSent(F.byteLength), P >= this.privRequestSession.bytesSent ? Q = 0 : Q = Math.max(0, C - Date.now())), Q !== 0 && await this.delay(Q), F !== null && (C = Date.now() + F.byteLength * 1e3 / (S.avgBytesPerSec * 2)), !this.privIsDisposed && !this.privRequestSession.isSpeechEnded && this.privRequestSession.isRecognizing && this.privRequestSession.recogNumber === A)
12698
+ if (w.send(new c.SpeechConnectionMessage(e.MessageType.Binary, "audio", this.privRequestSession.requestId, null, F)).catch(() => {
12699
12699
  this.privRequestSession.onServiceTurnEndResponse(this.privRecognizerConfig.isContinuousRecognition).catch(() => {
12700
12700
  });
12701
12701
  }), O != null && O.isEnd)
@@ -12714,16 +12714,16 @@ function tm() {
12714
12714
  let C = 0, y = "";
12715
12715
  for (; this.privRequestSession.numConnectionAttempts <= this.privRecognizerConfig.maxRetryCount; ) {
12716
12716
  this.privRequestSession.onRetryConnection();
12717
- const w = await (f ? this.privAuthentication.fetchOnExpiry(this.privAuthFetchEventId) : this.privAuthentication.fetch(this.privAuthFetchEventId));
12717
+ const A = await (f ? this.privAuthentication.fetchOnExpiry(this.privAuthFetchEventId) : this.privAuthentication.fetch(this.privAuthFetchEventId));
12718
12718
  await this.privRequestSession.onAuthCompleted(!1);
12719
- const I = await this.privConnectionFactory.create(this.privRecognizerConfig, w, this.privConnectionId);
12719
+ const I = await this.privConnectionFactory.create(this.privRecognizerConfig, A, this.privConnectionId);
12720
12720
  this.privRequestSession.listenForServiceTelemetry(I.events), I.events.attach((O) => {
12721
12721
  this.connectionEvents.onEvent(O);
12722
12722
  });
12723
- const A = await I.open();
12724
- if (A.statusCode === 200)
12725
- return await this.privRequestSession.onConnectionEstablishCompleted(A.statusCode), Promise.resolve(I);
12726
- A.statusCode === 1006 && (f = !0), C = A.statusCode, y = A.reason;
12723
+ const w = await I.open();
12724
+ if (w.statusCode === 200)
12725
+ return await this.privRequestSession.onConnectionEstablishCompleted(w.statusCode), Promise.resolve(I);
12726
+ w.statusCode === 1006 && (f = !0), C = w.statusCode, y = w.reason;
12727
12727
  }
12728
12728
  return await this.privRequestSession.onConnectionEstablishCompleted(C, y), Promise.reject(`Unable to contact server. StatusCode: ${C}, ${this.privRecognizerConfig.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Endpoint)} Reason: ${y}`);
12729
12729
  }
@@ -12978,8 +12978,8 @@ function sm() {
12978
12978
  let h = p.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Endpoint, void 0);
12979
12979
  const v = p.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Region, void 0), g = i.ConnectionFactoryBase.getHostSuffix(v), m = p.parameters.getProperty(r.PropertyId.SpeechServiceConnection_Host, "wss://" + v + ".stt.speech" + g), f = {}, S = p.parameters.getProperty(r.PropertyId.SpeechServiceConnection_EndpointId, void 0), C = p.parameters.getProperty(r.PropertyId.SpeechServiceConnection_RecoLanguage, void 0);
12980
12980
  if (S ? (!h || h.search(o.QueryParameterNames.CustomSpeechDeploymentId) === -1) && (f[o.QueryParameterNames.CustomSpeechDeploymentId] = S) : C && (!h || h.search(o.QueryParameterNames.Language) === -1) && (f[o.QueryParameterNames.Language] = C), (!h || h.search(o.QueryParameterNames.Format) === -1) && (f[o.QueryParameterNames.Format] = p.parameters.getProperty(e.OutputFormatPropertyName, r.OutputFormat[r.OutputFormat.Simple]).toLowerCase()), p.autoDetectSourceLanguages !== void 0 && (f[o.QueryParameterNames.EnableLanguageId] = "true"), this.setCommonUrlParams(p, f, h), h) {
12981
- const A = new URL(h), O = A.pathname;
12982
- (O === "" || O === "/") && (A.pathname = this.universalUri + p.recognitionEndpointVersion, h = await i.ConnectionFactoryBase.getRedirectUrlFromEndpoint(A.toString()));
12981
+ const w = new URL(h), O = w.pathname;
12982
+ (O === "" || O === "/") && (w.pathname = this.universalUri + p.recognitionEndpointVersion, h = await i.ConnectionFactoryBase.getRedirectUrlFromEndpoint(w.toString()));
12983
12983
  }
12984
12984
  if (!h)
12985
12985
  switch (p.recognitionMode) {
@@ -12995,8 +12995,8 @@ function sm() {
12995
12995
  }
12996
12996
  const y = {};
12997
12997
  d.token !== void 0 && d.token !== "" && (y[d.headerName] = d.token), y[a.HeaderNames.ConnectionId] = l, y.connectionId = l;
12998
- const P = p.parameters.getProperty("SPEECH-EnableWebsocketCompression", "false") === "true", w = new t.WebsocketConnection(h, f, y, new c.WebsocketMessageFormatter(), t.ProxyInfo.fromRecognizerConfig(p), P, l), I = w.uri;
12999
- return p.parameters.setProperty(r.PropertyId.SpeechServiceConnection_Url, I), w;
12998
+ const P = p.parameters.getProperty("SPEECH-EnableWebsocketCompression", "false") === "true", A = new t.WebsocketConnection(h, f, y, new c.WebsocketMessageFormatter(), t.ProxyInfo.fromRecognizerConfig(p), P, l), I = A.uri;
12999
+ return p.parameters.setProperty(r.PropertyId.SpeechServiceConnection_Url, I), A;
13000
13000
  }
13001
13001
  };
13002
13002
  return an.SpeechConnectionFactory = s, an;
@@ -13015,14 +13015,14 @@ function om() {
13015
13015
  let l = u.parameters.getProperty(e.PropertyId.SpeechServiceConnection_Endpoint, void 0);
13016
13016
  const h = u.parameters.getProperty(e.PropertyId.SpeechServiceConnection_Region, void 0), v = i.ConnectionFactoryBase.getHostSuffix(h), g = u.parameters.getProperty(e.PropertyId.SpeechServiceConnection_Host, "wss://" + h + ".stt.speech" + v), m = {}, f = u.parameters.getProperty(e.PropertyId.SpeechServiceConnection_EndpointId, void 0), S = u.parameters.getProperty(e.PropertyId.SpeechServiceConnection_RecoLanguage, void 0);
13017
13017
  if (f ? (!l || l.search(o.QueryParameterNames.CustomSpeechDeploymentId) === -1) && (m[o.QueryParameterNames.CustomSpeechDeploymentId] = f) : S && (!l || l.search(o.QueryParameterNames.Language) === -1) && (m[o.QueryParameterNames.Language] = S), u.autoDetectSourceLanguages !== void 0 && (m[o.QueryParameterNames.EnableLanguageId] = "true"), this.setV2UrlParams(u, m, l), l) {
13018
- const I = new URL(l), A = I.pathname;
13019
- (A === "" || A === "/") && (I.pathname = this.universalUri, l = await i.ConnectionFactoryBase.getRedirectUrlFromEndpoint(I.toString()));
13018
+ const I = new URL(l), w = I.pathname;
13019
+ (w === "" || w === "/") && (I.pathname = this.universalUri, l = await i.ConnectionFactoryBase.getRedirectUrlFromEndpoint(I.toString()));
13020
13020
  }
13021
13021
  l || (l = `${g}${this.universalUri}`);
13022
13022
  const C = {};
13023
13023
  p.token !== void 0 && p.token !== "" && (C[p.headerName] = p.token), C[a.HeaderNames.ConnectionId] = d;
13024
- const y = u.parameters.getProperty("SPEECH-EnableWebsocketCompression", "false") === "true", P = new t.WebsocketConnection(l, m, C, new c.WebsocketMessageFormatter(), t.ProxyInfo.fromRecognizerConfig(u), y, d), w = P.uri;
13025
- return u.parameters.setProperty(e.PropertyId.SpeechServiceConnection_Url, w), P;
13024
+ const y = u.parameters.getProperty("SPEECH-EnableWebsocketCompression", "false") === "true", P = new t.WebsocketConnection(l, m, C, new c.WebsocketMessageFormatter(), t.ProxyInfo.fromRecognizerConfig(u), y, d), A = P.uri;
13025
+ return u.parameters.setProperty(e.PropertyId.SpeechServiceConnection_Url, A), P;
13026
13026
  }
13027
13027
  setV2UrlParams(u, p, d) {
13028
13028
  (/* @__PURE__ */ new Map([
@@ -14704,19 +14704,19 @@ function Nm() {
14704
14704
  case "speech.keyword":
14705
14705
  const P = o.SpeechKeyword.fromJSON(h.textBody, this.privRequestSession.currentTurnAudioOffset);
14706
14706
  g = new c.SpeechRecognitionResult(this.privRequestSession.requestId, P.Status === "Accepted" ? c.ResultReason.RecognizedKeyword : c.ResultReason.NoMatch, P.Text, P.Duration, P.Offset, void 0, void 0, void 0, void 0, P.asJson(), v), P.Status !== "Accepted" && (this.privLastResult = g);
14707
- const w = new c.SpeechRecognitionEventArgs(g, g.duration, g.resultId);
14707
+ const A = new c.SpeechRecognitionEventArgs(g, g.duration, g.resultId);
14708
14708
  if (this.privDialogServiceConnector.recognized)
14709
14709
  try {
14710
- this.privDialogServiceConnector.recognized(this.privDialogServiceConnector, w);
14710
+ this.privDialogServiceConnector.recognized(this.privDialogServiceConnector, A);
14711
14711
  } catch {
14712
14712
  }
14713
14713
  m = !0;
14714
14714
  break;
14715
14715
  case "audio":
14716
14716
  {
14717
- const I = h.requestId.toUpperCase(), A = this.privTurnStateManager.GetTurn(I);
14717
+ const I = h.requestId.toUpperCase(), w = this.privTurnStateManager.GetTurn(I);
14718
14718
  try {
14719
- h.binaryBody ? A.audioStream.write(h.binaryBody) : A.endAudioStream();
14719
+ h.binaryBody ? w.audioStream.write(h.binaryBody) : w.endAudioStream();
14720
14720
  } catch {
14721
14721
  }
14722
14722
  }
@@ -14774,13 +14774,13 @@ function Nm() {
14774
14774
  await this.privRequestSession.onAudioSourceAttachCompleted(P, !1), this.privRecognizerConfig.SpeechServiceConfig.Context.audio = { source: y };
14775
14775
  try {
14776
14776
  await m, await f;
14777
- } catch (A) {
14778
- return await this.cancelRecognition(this.privRequestSession.sessionId, this.privRequestSession.requestId, c.CancellationReason.Error, c.CancellationErrorCode.ConnectionFailure, A), Promise.resolve();
14777
+ } catch (w) {
14778
+ return await this.cancelRecognition(this.privRequestSession.sessionId, this.privRequestSession.requestId, c.CancellationReason.Error, c.CancellationErrorCode.ConnectionFailure, w), Promise.resolve();
14779
14779
  }
14780
- const w = new c.SessionEventArgs(this.privRequestSession.sessionId);
14781
- this.privRecognizer.sessionStarted && this.privRecognizer.sessionStarted(this.privRecognizer, w), this.sendAudio(P).then(() => {
14782
- }, async (A) => {
14783
- await this.cancelRecognition(this.privRequestSession.sessionId, this.privRequestSession.requestId, c.CancellationReason.Error, c.CancellationErrorCode.RuntimeError, A);
14780
+ const A = new c.SessionEventArgs(this.privRequestSession.sessionId);
14781
+ this.privRecognizer.sessionStarted && this.privRecognizer.sessionStarted(this.privRecognizer, A), this.sendAudio(P).then(() => {
14782
+ }, async (w) => {
14783
+ await this.cancelRecognition(this.privRequestSession.sessionId, this.privRequestSession.requestId, c.CancellationReason.Error, c.CancellationErrorCode.RuntimeError, w);
14784
14784
  });
14785
14785
  }
14786
14786
  // Establishes a websocket connection to the end point.
@@ -14807,15 +14807,15 @@ function Nm() {
14807
14807
  }
14808
14808
  break;
14809
14809
  case "speech.startdetected":
14810
- const P = o.SpeechDetected.fromJSON(C.textBody, this.privRequestSession.currentTurnAudioOffset), w = new c.RecognitionEventArgs(P.Offset, this.privRequestSession.sessionId);
14811
- this.privRecognizer.speechStartDetected && this.privRecognizer.speechStartDetected(this.privRecognizer, w);
14810
+ const P = o.SpeechDetected.fromJSON(C.textBody, this.privRequestSession.currentTurnAudioOffset), A = new c.RecognitionEventArgs(P.Offset, this.privRequestSession.sessionId);
14811
+ this.privRecognizer.speechStartDetected && this.privRecognizer.speechStartDetected(this.privRecognizer, A);
14812
14812
  break;
14813
14813
  case "speech.enddetected":
14814
14814
  let I;
14815
14815
  C.textBody.length > 0 ? I = C.textBody : I = "{ Offset: 0 }";
14816
- const A = o.SpeechDetected.fromJSON(I, this.privRequestSession.currentTurnAudioOffset);
14817
- this.privRequestSession.onServiceRecognized(A.Offset);
14818
- const O = new c.RecognitionEventArgs(A.Offset, this.privRequestSession.sessionId);
14816
+ const w = o.SpeechDetected.fromJSON(I, this.privRequestSession.currentTurnAudioOffset);
14817
+ this.privRequestSession.onServiceRecognized(w.Offset);
14818
+ const O = new c.RecognitionEventArgs(w.Offset, this.privRequestSession.sessionId);
14819
14819
  this.privRecognizer.speechEndDetected && this.privRecognizer.speechEndDetected(this.privRecognizer, O);
14820
14820
  break;
14821
14821
  case "turn.end":
@@ -15029,25 +15029,25 @@ function Lm() {
15029
15029
  C.headers = S, this.privRestAdapter.options = C;
15030
15030
  const y = `https://${l}${this.privRestPath}`;
15031
15031
  this.privRestAdapter.request(t.RestRequestType.Post, y, f, null).then((P) => {
15032
- const w = t.RestMessageAdapter.extractHeaderValue(this.privRequestParams.requestId, P.headers);
15032
+ const A = t.RestMessageAdapter.extractHeaderValue(this.privRequestParams.requestId, P.headers);
15033
15033
  if (!P.ok) {
15034
15034
  if (u) {
15035
- let A = this.privErrors.invalidCreateJoinConversationResponse.replace("{status}", P.status.toString()), O;
15035
+ let w = this.privErrors.invalidCreateJoinConversationResponse.replace("{status}", P.status.toString()), O;
15036
15036
  try {
15037
- O = JSON.parse(P.data), A += ` [${O.error.code}: ${O.error.message}]`;
15037
+ O = JSON.parse(P.data), w += ` [${O.error.code}: ${O.error.message}]`;
15038
15038
  } catch {
15039
- A += ` [${P.data}]`;
15039
+ w += ` [${P.data}]`;
15040
15040
  }
15041
- w && (A += ` ${w}`), u(A);
15041
+ A && (w += ` ${A}`), u(w);
15042
15042
  }
15043
15043
  return;
15044
15044
  }
15045
15045
  const I = JSON.parse(P.data);
15046
- if (I && (I.requestId = w), s) {
15046
+ if (I && (I.requestId = A), s) {
15047
15047
  try {
15048
15048
  s(I);
15049
- } catch (A) {
15050
- u && u(A);
15049
+ } catch (w) {
15050
+ u && u(w);
15051
15051
  }
15052
15052
  s = void 0;
15053
15053
  }
@@ -15655,7 +15655,7 @@ function Gm() {
15655
15655
  const m = n.CommandResponsePayload.fromJSON(l.textBody);
15656
15656
  switch (m.command.toLowerCase()) {
15657
15657
  case "participantlist":
15658
- const w = n.ParticipantsListPayloadResponse.fromJSON(l.textBody), I = w.participants.map((K) => ({
15658
+ const A = n.ParticipantsListPayloadResponse.fromJSON(l.textBody), I = A.participants.map((K) => ({
15659
15659
  avatar: K.avatar,
15660
15660
  displayName: K.nickname,
15661
15661
  id: K.participantId,
@@ -15664,7 +15664,7 @@ function Gm() {
15664
15664
  isUsingTts: K.usetts,
15665
15665
  preferredLanguage: K.locale
15666
15666
  }));
15667
- this.privConversationServiceConnector.participantsListReceived && this.privConversationServiceConnector.participantsListReceived(this.privConversationServiceConnector, new a.ParticipantsListEventArgs(w.roomid, w.token, w.translateTo, w.profanityFilter, w.roomProfanityFilter, w.roomLocked, w.muteAll, I, h));
15667
+ this.privConversationServiceConnector.participantsListReceived && this.privConversationServiceConnector.participantsListReceived(this.privConversationServiceConnector, new a.ParticipantsListEventArgs(A.roomid, A.token, A.translateTo, A.profanityFilter, A.roomProfanityFilter, A.roomLocked, A.muteAll, I, h));
15668
15668
  break;
15669
15669
  case "settranslatetolanguages":
15670
15670
  this.privConversationServiceConnector.participantUpdateCommandReceived && this.privConversationServiceConnector.participantUpdateCommandReceived(this.privConversationServiceConnector, new a.ParticipantAttributeEventArgs(m.participantId, o.ConversationTranslatorCommandTypes.setTranslateToLanguages, m.value, h));
@@ -15691,14 +15691,14 @@ function Gm() {
15691
15691
  this.privConversationServiceConnector.participantUpdateCommandReceived && this.privConversationServiceConnector.participantUpdateCommandReceived(this.privConversationServiceConnector, new a.ParticipantAttributeEventArgs(m.participantId, o.ConversationTranslatorCommandTypes.changeNickname, m.value, h));
15692
15692
  break;
15693
15693
  case "joinsession":
15694
- const A = n.ParticipantPayloadResponse.fromJSON(l.textBody), O = {
15695
- avatar: A.avatar,
15696
- displayName: A.nickname,
15697
- id: A.participantId,
15698
- isHost: A.ishost,
15699
- isMuted: A.ismuted,
15700
- isUsingTts: A.usetts,
15701
- preferredLanguage: A.locale
15694
+ const w = n.ParticipantPayloadResponse.fromJSON(l.textBody), O = {
15695
+ avatar: w.avatar,
15696
+ displayName: w.nickname,
15697
+ id: w.participantId,
15698
+ isHost: w.ishost,
15699
+ isMuted: w.ismuted,
15700
+ isUsingTts: w.usetts,
15701
+ preferredLanguage: w.locale
15702
15702
  };
15703
15703
  this.privConversationServiceConnector.participantJoinCommandReceived && this.privConversationServiceConnector.participantJoinCommandReceived(this.privConversationServiceConnector, new a.ParticipantEventArgs(O, h));
15704
15704
  break;
@@ -17447,28 +17447,30 @@ const _S = { class: "max-h-screen overflow-auto md:p-6 p-3 flex flex-col items-c
17447
17447
  This is a role-play simulation designed to mirror real-world scenarios. Just follow the prompts, speak naturally, and let the AI guide you. Based on your responses, you’ll receive personalized feedback to help sharpen your client engagement skills.`, LS = /* @__PURE__ */ Pe({
17448
17448
  __name: "greetingsSimulation",
17449
17449
  setup(t) {
17450
- const e = be(), r = St(), i = M(!1), c = M(!1), a = M(!0), o = `Hi, get ready to test your ${e.skillNameForSimulation || "GenpactNext"} knowledge! Here’s what to expect:`, n = ` Hi ${e.learnerName || "Expert"}, and welcome!`, s = ` Hi, ${e.learnerName || "Expert"}. Get ready to test your ${e.skillNameForSimulation || "GenpactNext"} knowledge! Here’s what to expect:`, u = Lo(() => ["32b7d6d8-f7a8-40a0-ab84-8784f25897ef"].includes(e.simulationSkill) ? o : ["7bd03cb6-79ad-4e2a-8a8b-88dc16147369"].includes(e.simulationSkill) ? n : s), p = `Hi, get ready to test your ${e.skillNameForSimulation || "GenpactNext"} knowledge! Here’s what to expect:`, d = ` AI Guru will ask you a few questions on '${e.skillNameForSimulation.trim() || "GenpactNext"} '. Please provide detailed and thoughtful answers to the best of your knowledge. At the end, you'll receive a feedback, outlining your strengths and improvement areas.`, l = Lo(() => ["32b7d6d8-f7a8-40a0-ab84-8784f25897ef"].includes(e.simulationSkill) ? p : ["7bd03cb6-79ad-4e2a-8a8b-88dc16147369"].includes(e.simulationSkill) ? jS : d), h = async () => {
17451
- var v, g, m, f, S, C;
17450
+ const e = be(), r = St(), i = M(!1), c = M(!1), a = M(!0), o = () => {
17451
+ e.simulationSkill === "7bd03cb6-79ad-4e2a-8a8b-88dc16147369" ? (e.beginFlag = !1, i.value = !0) : (e.beginFlag = !1, v());
17452
+ }, n = `Hi, get ready to test your ${e.skillNameForSimulation || "GenpactNext"} knowledge! Here’s what to expect:`, s = ` Hi ${e.learnerName || "Expert"}, and welcome!`, u = ` Hi, ${e.learnerName || "Expert"}. Get ready to test your ${e.skillNameForSimulation || "GenpactNext"} knowledge! Here’s what to expect:`, p = Lo(() => ["32b7d6d8-f7a8-40a0-ab84-8784f25897ef"].includes(e.simulationSkill) ? n : ["7bd03cb6-79ad-4e2a-8a8b-88dc16147369"].includes(e.simulationSkill) ? s : u), d = `Hi, get ready to test your ${e.skillNameForSimulation || "GenpactNext"} knowledge! Here’s what to expect:`, l = ` AI Guru will ask you a few questions on '${e.skillNameForSimulation.trim() || "GenpactNext"} '. Please provide detailed and thoughtful answers to the best of your knowledge. At the end, you'll receive a feedback, outlining your strengths and improvement areas.`, h = Lo(() => ["32b7d6d8-f7a8-40a0-ab84-8784f25897ef"].includes(e.simulationSkill) ? d : ["7bd03cb6-79ad-4e2a-8a8b-88dc16147369"].includes(e.simulationSkill) ? jS : l), v = async () => {
17453
+ var g, m, f, S, C, y;
17452
17454
  console.log("firstQuestion fired==>"), a.value = !1;
17453
17455
  try {
17454
- const y = {
17456
+ const P = {
17455
17457
  question: "Hello, let's start"
17456
17458
  };
17457
- if (e.conversations.push(y), e.assessmentQuestion) {
17458
- const P = {
17459
+ if (e.conversations.push(P), e.assessmentQuestion) {
17460
+ const A = {
17459
17461
  past_messages: e.past_messages,
17460
17462
  question: "Let's begin",
17461
17463
  bandCode: e.band_code,
17462
17464
  userGenomeFunction: e.genomeFunctionValue,
17463
17465
  userGenomeRole: e.genomeRoleValue,
17464
17466
  skillId: e.simulationSkill || "937e84ef-e95d-4327-9afe-e7be2bf420eb"
17465
- }, w = M("");
17467
+ }, I = M("");
17466
17468
  e.dotLoader = !0, le.value = "listening";
17467
- const I = await gt("/sqlTest", P);
17468
- e.addConversation((v = I.data) == null ? void 0 : v.body, "new"), w.value = (m = (g = I == null ? void 0 : I.data) == null ? void 0 : g.body) == null ? void 0 : m.answer, e.dotLoader = !1, le.value = "bot", e.showVoiceComponent && (e.elevenLabs ? await Qd(w.value.replace(/[*#]/g, "")) : await Jd(w.value.replace(/[*#]/g, "")));
17469
+ const w = await gt("/sqlTest", A);
17470
+ e.addConversation((g = w.data) == null ? void 0 : g.body, "new"), I.value = (f = (m = w == null ? void 0 : w.data) == null ? void 0 : m.body) == null ? void 0 : f.answer, e.dotLoader = !1, le.value = "bot", e.showVoiceComponent && (e.elevenLabs ? await Qd(I.value.replace(/[*#]/g, "")) : await Jd(I.value.replace(/[*#]/g, "")));
17469
17471
  } else {
17470
17472
  c.value = !0, e.dotLoader = !0, console.log("firstQuestion fired==> else part");
17471
- const P = {
17473
+ const A = {
17472
17474
  assessmentId: e.simulationSkill,
17473
17475
  question: "Hello, Let's begin",
17474
17476
  past_messages: e.past_messages,
@@ -17481,28 +17483,28 @@ This is a role-play simulation designed to mirror real-world scenarios. Just fol
17481
17483
  storeFinalFeedback: e.storeFinalFeedback,
17482
17484
  storeFinalScore: e.storeFinalScore
17483
17485
  };
17484
- let w;
17485
- Dt === "prod" ? w = "/ManualExportActivity" : w = "/fetchCompletionDemo";
17486
- const I = await gt(w, P);
17487
- e.addConversation((f = I.data) == null ? void 0 : f.body, "assessment"), e.dotLoader = !1;
17486
+ let I;
17487
+ Dt === "prod" ? I = "/ManualExportActivity" : I = "/fetchCompletionDemo";
17488
+ const w = await gt(I, A);
17489
+ e.addConversation((S = w.data) == null ? void 0 : S.body, "assessment"), e.dotLoader = !1;
17488
17490
  }
17489
- } catch (y) {
17490
- if (console.error(y), Ce.isAxiosError(y)) {
17491
- const P = y;
17492
- console.log("Error found!"), console.error((S = P.response) == null ? void 0 : S.data), r.setError(((C = P.response) == null ? void 0 : C.data).message);
17491
+ } catch (P) {
17492
+ if (console.error(P), Ce.isAxiosError(P)) {
17493
+ const A = P;
17494
+ console.log("Error found!"), console.error((C = A.response) == null ? void 0 : C.data), r.setError(((y = A.response) == null ? void 0 : y.data).message);
17493
17495
  } else
17494
17496
  r.setError("There was an issue, please contact helpmate");
17495
17497
  }
17496
17498
  };
17497
- return (v, g) => (N(), j("section", null, [
17499
+ return (g, m) => (N(), j("section", null, [
17498
17500
  ge(sp, { "is-visible": i.value }, {
17499
17501
  default: op(() => [
17500
17502
  R("div", _S, [
17501
17503
  TS,
17502
17504
  bS,
17503
17505
  R("button", {
17504
- onClick: g[0] || (g[0] = (m) => {
17505
- i.value = !1, h();
17506
+ onClick: m[0] || (m[0] = (f) => {
17507
+ i.value = !1, v();
17506
17508
  }),
17507
17509
  class: "bg-[#021c42] rounded-md px-4 py-1.5 text-center text-[16px] text-white cursor-pointer italic hover:bg-blue-700"
17508
17510
  }, " Proceed ")
@@ -17512,17 +17514,15 @@ This is a role-play simulation designed to mirror real-world scenarios. Just fol
17512
17514
  }, 8, ["is-visible"]),
17513
17515
  R("div", MS, [
17514
17516
  R("div", OS, [
17515
- R("div", DS, Ve(u.value), 1),
17516
- R("div", kS, Ve(l.value), 1)
17517
+ R("div", DS, Ve(p.value), 1),
17518
+ R("div", kS, Ve(h.value), 1)
17517
17519
  ])
17518
17520
  ]),
17519
17521
  k(e).beginFlag ? (N(), j("span", NS, [
17520
17522
  R("button", {
17521
17523
  type: "button",
17522
17524
  id: "button",
17523
- onClick: g[1] || (g[1] = (m) => {
17524
- k(e).beginFlag = !1, i.value = !0;
17525
- }),
17525
+ onClick: o,
17526
17526
  class: "mt-4 bg-[#021c42] rounded-md px-4 py-1.5 text-center text-[16px] text-white cursor-pointer italic hover:bg-blue-700"
17527
17527
  }, " Let's begin ")
17528
17528
  ])) : _e("", !0),
@@ -18027,7 +18027,7 @@ const _C = /* @__PURE__ */ Ft(IC, [["__scopeId", "data-v-e55a322f"]]), TC = { cl
18027
18027
  console.log("ques", y), e.conversations.push(y);
18028
18028
  const P = M("");
18029
18029
  P.value = e.question || e.recognizedText, e.question = "";
18030
- const w = {
18030
+ const A = {
18031
18031
  ohr_id: i.value,
18032
18032
  past_messages: e.past_messages,
18033
18033
  question: P.value || e.recognizedText,
@@ -18035,10 +18035,10 @@ const _C = /* @__PURE__ */ Ft(IC, [["__scopeId", "data-v-e55a322f"]]), TC = { cl
18035
18035
  userGenomeFunction: e.genomeFunctionValue,
18036
18036
  userGenomeRole: e.genomeRoleValue,
18037
18037
  skillId: e.simulationSkill || "937e84ef-e95d-4327-9afe-e7be2bf420eb"
18038
- }, I = await gt("/sqlTest", w);
18038
+ }, I = await gt("/sqlTest", A);
18039
18039
  if (i.value = (h = (l = I.data) == null ? void 0 : l.body) == null ? void 0 : h.ohr, ((g = (v = I == null ? void 0 : I.data) == null ? void 0 : v.body) == null ? void 0 : g.answer) === "") {
18040
18040
  console.log("in the if case calling api again");
18041
- const O = await gt("/sqlTest", w);
18041
+ const O = await gt("/sqlTest", A);
18042
18042
  e.addConversation((m = O.data) == null ? void 0 : m.body, "new");
18043
18043
  } else
18044
18044
  console.log("in the else case"), e.addConversation((f = I.data) == null ? void 0 : f.body, "new");
@@ -18121,21 +18121,21 @@ const _C = /* @__PURE__ */ Ft(IC, [["__scopeId", "data-v-e55a322f"]]), TC = { cl
18121
18121
  ]));
18122
18122
  }
18123
18123
  });
18124
- const Cy = /* @__PURE__ */ Ft(Sy, [["__scopeId", "data-v-e951c0ed"]]), po = (t) => (Wn("data-v-b13dca42"), t = t(), Vn(), t), yy = { class: "z-3 absolute bottom-0 mt-4 w-full bg-white px-2 md:px-6" }, Py = { class: "py-2" }, Ry = { class: "flex" }, Ey = { class: "z-1 relative w-full" }, wy = { class: "-text-blue absolute inset-y-0 bottom-8 right-0 flex cursor-pointer items-center justify-center pr-4 md:bottom-5 md:pr-8 lg:pr-8" }, Ay = { key: 0 }, Iy = ["disabled"], _y = /* @__PURE__ */ po(() => /* @__PURE__ */ R("path", { d: "M15.854.146a.5.5 0 0 1 .11.54l-5.819 14.547a.75.75 0 0 1-1.329.124l-3.178-4.995L.643 7.184a.75.75 0 0 1 .124-1.33L15.314.037a.5.5 0 0 1 .54.11ZM6.636 10.07l2.761 4.338L14.13 2.576 6.636 10.07Zm6.787-8.201L1.591 6.602l4.339 2.76 7.494-7.493Z" }, null, -1)), Ty = [
18125
- _y
18126
- ], by = {
18124
+ const Cy = /* @__PURE__ */ Ft(Sy, [["__scopeId", "data-v-e951c0ed"]]), po = (t) => (Wn("data-v-bbd70acc"), t = t(), Vn(), t), yy = { class: "z-3 absolute bottom-0 mt-4 w-full bg-white px-2 md:px-6" }, Py = { class: "py-2" }, Ry = { class: "flex" }, Ey = { class: "z-1 relative w-full" }, wy = { class: "relative w-full" }, Ay = { class: "absolute inset-y-0 right-0 flex items-center md:pr-8 pr-3" }, Iy = { key: 0 }, _y = ["disabled"], Ty = /* @__PURE__ */ po(() => /* @__PURE__ */ R("path", { d: "M15.854.146a.5.5 0 0 1 .11.54l-5.819 14.547a.75.75 0 0 1-1.329.124l-3.178-4.995L.643 7.184a.75.75 0 0 1 .124-1.33L15.314.037a.5.5 0 0 1 .54.11ZM6.636 10.07l2.761 4.338L14.13 2.576 6.636 10.07Zm6.787-8.201L1.591 6.602l4.339 2.76 7.494-7.493Z" }, null, -1)), by = [
18125
+ Ty
18126
+ ], My = {
18127
18127
  key: 1,
18128
18128
  class: "tooltip"
18129
- }, My = /* @__PURE__ */ po(() => /* @__PURE__ */ R("span", { class: "tooltiptext fontFormattingAMA" }, "Waiting for response", -1)), Oy = /* @__PURE__ */ po(() => /* @__PURE__ */ R("img", {
18129
+ }, Oy = /* @__PURE__ */ po(() => /* @__PURE__ */ R("span", { class: "tooltiptext fontFormattingAMA" }, "Waiting for response", -1)), Dy = /* @__PURE__ */ po(() => /* @__PURE__ */ R("img", {
18130
+ class: "w-7 h-7",
18130
18131
  src: ip,
18131
- alt: "watingIcon",
18132
- class: "ml-2"
18133
- }, null, -1)), Dy = [
18134
- My,
18135
- Oy
18136
- ], ky = { class: "flex justify-start gap-2 md:gap-1" }, Ny = ["disabled"], zy = /* @__PURE__ */ po(() => /* @__PURE__ */ R("div", { class: "pt-1 text-start font-sans text-xs font-bold text-gray-500 md:text-sm" }, [
18132
+ alt: "watingIcon"
18133
+ }, null, -1)), ky = [
18134
+ Oy,
18135
+ Dy
18136
+ ], Ny = { class: "flex justify-start gap-2 md:gap-1" }, zy = ["disabled"], jy = /* @__PURE__ */ po(() => /* @__PURE__ */ R("div", { class: "pt-1 text-start font-sans text-xs font-bold text-gray-500 md:text-sm" }, [
18137
18137
  /* @__PURE__ */ R("div", null, " Note: AI-generated response may not be fully accurate; check with experts for critical information. ")
18138
- ], -1)), jy = /* @__PURE__ */ Pe({
18138
+ ], -1)), Ly = /* @__PURE__ */ Pe({
18139
18139
  __name: "NewMDMQuestionSimulation",
18140
18140
  setup(t) {
18141
18141
  const e = be(), r = St(), i = M("");
@@ -18193,61 +18193,62 @@ const Cy = /* @__PURE__ */ Ft(Sy, [["__scopeId", "data-v-e951c0ed"]]), po = (t)
18193
18193
  R("div", Ry, [
18194
18194
  R("div", Ey, [
18195
18195
  R("div", wy, [
18196
- n.value ? (N(), j("div", by, Dy)) : (N(), j("div", Ay, [
18197
- (N(), j("svg", {
18198
- disabled: k(e).beginFlag === !0,
18196
+ R("div", Ay, [
18197
+ n.value ? (N(), j("div", My, ky)) : (N(), j("div", Iy, [
18198
+ (N(), j("svg", {
18199
+ disabled: k(e).beginFlag === !0,
18200
+ class: kt([
18201
+ k(e).beginFlag === !0 ? "cursor-not-allowed text-gray-600" : "cursor-pointer",
18202
+ "bi bi-send ml-2 h-7 w-7 rotate-45 rounded"
18203
+ ]),
18204
+ onClick: d[0] || (d[0] = (l) => k(e).searchBoxdisable === !1 || k(e).beginFlag === !0 ? "" : u()),
18205
+ xmlns: "http://www.w3.org/2000/svg",
18206
+ fill: "currentColor",
18207
+ viewBox: "0 0 16 16"
18208
+ }, by, 10, _y))
18209
+ ]))
18210
+ ]),
18211
+ R("div", Ny, [
18212
+ Jo(R("input", {
18213
+ onKeyup: [
18214
+ d[1] || (d[1] = cp((l) => n.value ? "" : u(), ["enter"])),
18215
+ d[3] || (d[3] = (l) => s(k(e).question))
18216
+ ],
18217
+ disabled: k(e).beginFlag === !0 || k(e).searchBoxdisable === !1,
18199
18218
  class: kt([
18200
- k(e).beginFlag === !0 ? "cursor-not-allowed text-gray-600" : "pointer -text-blue",
18201
- "bi bi-send ml-2 h-7 w-7 rotate-45 rounded"
18219
+ k(e).beginFlag === !0 || k(e).searchBoxdisable === !1 ? "cursor-not-allowed bg-[#e5e7eb]" : "bg-white",
18220
+ "fontFormattingAMA rounded-lg flex w-full border border-genpact-darkest-teal py-[18px] pl-[30px] pr-[70px]"
18202
18221
  ]),
18203
- onClick: d[0] || (d[0] = (l) => k(e).searchBoxdisable === !1 || k(e).beginFlag === !0 ? "" : u()),
18204
- xmlns: "http://www.w3.org/2000/svg",
18205
- fill: "currentColor",
18206
- viewBox: "0 0 16 16"
18207
- }, Ty, 10, Iy))
18208
- ]))
18209
- ]),
18210
- R("div", ky, [
18211
- Jo(R("textarea", {
18212
- rows: "2",
18213
- onKeyup: [
18214
- d[1] || (d[1] = cp((l) => u(), ["enter"])),
18215
- d[3] || (d[3] = (l) => s(k(e).question))
18216
- ],
18217
- disabled: k(e).beginFlag === !0,
18218
- class: kt([
18219
- k(e).beginFlag === !0 || k(e).searchBoxdisable === !1 ? "cursor-not-allowed bg-[#e5e7eb]" : "",
18220
- "fontFormattingAMA flex w-full border border-genpact-darkest-teal py-2 pl-2 pr-16 focus:outline-none sm:text-[15px] md:py-3 md:pl-6 lg:py-3 lg:pl-6"
18221
- ]),
18222
- "onUpdate:modelValue": d[2] || (d[2] = (l) => k(e).question = l),
18223
- placeholder: "Write your answer here...",
18224
- type: "text",
18225
- ref_key: "AiSearch",
18226
- ref: c
18227
- }, null, 42, Ny), [
18228
- [ap, k(e).question]
18222
+ "onUpdate:modelValue": d[2] || (d[2] = (l) => k(e).question = l),
18223
+ placeholder: "Write your answer here...",
18224
+ type: "text",
18225
+ ref_key: "AiSearch",
18226
+ ref: c
18227
+ }, null, 42, zy), [
18228
+ [ap, k(e).question]
18229
+ ])
18229
18230
  ])
18230
18231
  ]),
18231
- zy
18232
+ jy
18232
18233
  ])
18233
18234
  ])
18234
18235
  ])
18235
18236
  ]));
18236
18237
  }
18237
18238
  });
18238
- const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t) => (Wn("data-v-904e6953"), t = t(), Vn(), t), xy = /* @__PURE__ */ Xd(() => /* @__PURE__ */ R("div", { class: "max-h-screen overflow-auto md:p-6 p-3 flex flex-col items-center justify-center gap-4 text-[#161916]" }, [
18239
+ const xy = /* @__PURE__ */ Ft(Ly, [["__scopeId", "data-v-bbd70acc"]]), Xd = (t) => (Wn("data-v-904e6953"), t = t(), Vn(), t), qy = /* @__PURE__ */ Xd(() => /* @__PURE__ */ R("div", { class: "max-h-screen overflow-auto md:p-6 p-3 flex flex-col items-center justify-center gap-4 text-[#161916]" }, [
18239
18240
  /* @__PURE__ */ R("div", { class: "text-2xl font-[700]" }, "Assessment Access Required"),
18240
18241
  /* @__PURE__ */ R("h3", { className: "text-lg font-medium my-6 text-center text-gray-500" }, " You don't have access to this assessment. Please contact to administrator ")
18241
- ], -1)), qy = {
18242
+ ], -1)), By = {
18242
18243
  key: 1,
18243
18244
  className: "h-screen flex flex-col items-stretch"
18244
- }, By = { class: "z-20" }, Fy = /* @__PURE__ */ Xd(() => /* @__PURE__ */ R("div", { class: "mt-6" }, null, -1)), Uy = { class: "flex h-full items-stretch overflow-hidden" }, Hy = { class: "mx-auto w-11/12 bg-white" }, Wy = { class: "relative flex h-full w-full flex-col items-stretch overflow-hidden" }, Vy = {
18245
+ }, Fy = { class: "z-20" }, Uy = /* @__PURE__ */ Xd(() => /* @__PURE__ */ R("div", { class: "mt-6" }, null, -1)), Hy = { class: "flex h-full items-stretch overflow-hidden" }, Wy = { class: "mx-auto w-11/12 bg-white" }, Vy = { class: "relative flex h-full w-full flex-col items-stretch overflow-hidden" }, $y = {
18245
18246
  key: 0,
18246
18247
  class: "relative"
18247
- }, $y = {
18248
+ }, Ky = {
18248
18249
  key: 1,
18249
18250
  class: "relative"
18250
- }, Ky = /* @__PURE__ */ Pe({
18251
+ }, Jy = /* @__PURE__ */ Pe({
18251
18252
  __name: "AssessmentSimulation",
18252
18253
  setup(t) {
18253
18254
  const e = M(!0), r = M(!1), i = M(!1);
@@ -18261,12 +18262,12 @@ const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t)
18261
18262
  console.log("uuid. value", g), e.value = !0, await S(), await C(), c.assessmentQuestion || (console.log("getAssessmentDetails========================>"), await f()), e.value = !1;
18262
18263
  });
18263
18264
  const f = async () => {
18264
- var y, P, w, I, A, O, F, Q, G, K, $t, ii, si, oi, ai, ci, ui, pi, di, li, hi, vi, fi, gi, mi, Si, Ci, yi, Pi, Ri, Ei, wi, Ai, Ii, _i, Ti, bi, Kt, Jt, Mi, Oi, Di, ki, Ni, zi, ji, Li, Pt;
18265
+ var y, P, A, I, w, O, F, Q, G, K, $t, ii, si, oi, ai, ci, ui, pi, di, li, hi, vi, fi, gi, mi, Si, Ci, yi, Pi, Ri, Ei, wi, Ai, Ii, _i, Ti, bi, Kt, Jt, Mi, Oi, Di, ki, Ni, zi, ji, Li, Pt;
18265
18266
  console.log("Start getAssessmentDetails");
18266
18267
  try {
18267
18268
  Dt === "prod" ? p = `https://api-v2.genpact.com/api/my-profile/assessment-master/getAssessmentDetailsById?id=${c.simulationSkill}` : p = `https://api-v2-dev.genpact.com/api/my-profile/assessment-master/getAssessmentDetailsById?id=${c.simulationSkill}`;
18268
18269
  const _ = await xn(p);
18269
- o.value = (P = (y = _ == null ? void 0 : _.data) == null ? void 0 : y.data) == null ? void 0 : P.name, c.skillNameForSimulation = o.value, c.generatedPrompt = (I = (w = _ == null ? void 0 : _.data) == null ? void 0 : w.data) == null ? void 0 : I.generatedPrompt, c.assessmentType = (O = (A = _ == null ? void 0 : _.data) == null ? void 0 : A.data) == null ? void 0 : O.assessmentType, s.value = (Q = (F = _ == null ? void 0 : _.data) == null ? void 0 : F.data) == null ? void 0 : Q.status, h.value = (K = (G = _ == null ? void 0 : _.data) == null ? void 0 : G.data) == null ? void 0 : K.editMode, c.empSeniority = (ii = ($t = _ == null ? void 0 : _.data) == null ? void 0 : $t.data) == null ? void 0 : ii.empSeniority, c.functionRole = (oi = (si = _ == null ? void 0 : _.data) == null ? void 0 : si.data) == null ? void 0 : oi.functionRole, d = (ci = (ai = _ == null ? void 0 : _.data) == null ? void 0 : ai.data) == null ? void 0 : ci.accessBand, l = (pi = (ui = _ == null ? void 0 : _.data) == null ? void 0 : ui.data) == null ? void 0 : pi.accessOhrList;
18270
+ o.value = (P = (y = _ == null ? void 0 : _.data) == null ? void 0 : y.data) == null ? void 0 : P.name, c.skillNameForSimulation = o.value, c.generatedPrompt = (I = (A = _ == null ? void 0 : _.data) == null ? void 0 : A.data) == null ? void 0 : I.generatedPrompt, c.assessmentType = (O = (w = _ == null ? void 0 : _.data) == null ? void 0 : w.data) == null ? void 0 : O.assessmentType, s.value = (Q = (F = _ == null ? void 0 : _.data) == null ? void 0 : F.data) == null ? void 0 : Q.status, h.value = (K = (G = _ == null ? void 0 : _.data) == null ? void 0 : G.data) == null ? void 0 : K.editMode, c.empSeniority = (ii = ($t = _ == null ? void 0 : _.data) == null ? void 0 : $t.data) == null ? void 0 : ii.empSeniority, c.functionRole = (oi = (si = _ == null ? void 0 : _.data) == null ? void 0 : si.data) == null ? void 0 : oi.functionRole, d = (ci = (ai = _ == null ? void 0 : _.data) == null ? void 0 : ai.data) == null ? void 0 : ci.accessBand, l = (pi = (ui = _ == null ? void 0 : _.data) == null ? void 0 : ui.data) == null ? void 0 : pi.accessOhrList;
18270
18271
  const Le = (li = (di = _ == null ? void 0 : _.data) == null ? void 0 : di.data) == null ? void 0 : li.createdBy;
18271
18272
  if (c.mcqType = (vi = (hi = _ == null ? void 0 : _.data) == null ? void 0 : hi.data) == null ? void 0 : vi.mcqTypes, c.storeFinalFeedback = (gi = (fi = _ == null ? void 0 : _.data) == null ? void 0 : fi.data) == null ? void 0 : gi.storeFinalFeedback, c.storeFinalScore = (Si = (mi = _ == null ? void 0 : _.data) == null ? void 0 : mi.data) == null ? void 0 : Si.storeFinalScore, s.value === "inactive")
18272
18273
  r.value = !0;
@@ -18318,7 +18319,7 @@ const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t)
18318
18319
  a.setError("There was an issue, please contact helpmate");
18319
18320
  }
18320
18321
  }, S = async () => {
18321
- var y, P, w, I, A;
18322
+ var y, P, A, I, w;
18322
18323
  try {
18323
18324
  if (g === null)
18324
18325
  throw {
@@ -18339,16 +18340,16 @@ const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t)
18339
18340
  } catch (O) {
18340
18341
  if (console.error(O), e.value = !1, Ce.isAxiosError(O) && ((y = O.response) != null && y.status) || O && typeof O == "object" && "response" in O && O.response && typeof O.response == "object" && "status" in O.response && O.response.status) {
18341
18342
  const F = O;
18342
- if (console.log("Error found!"), console.error((P = F.response) == null ? void 0 : P.data), ((w = F.response) == null ? void 0 : w.status) === 424 || ((I = F.response) == null ? void 0 : I.status) === 425) {
18343
+ if (console.log("Error found!"), console.error((P = F.response) == null ? void 0 : P.data), ((A = F.response) == null ? void 0 : A.status) === 424 || ((I = F.response) == null ? void 0 : I.status) === 425) {
18343
18344
  i.value = !0;
18344
18345
  return;
18345
18346
  }
18346
- a.setError(((A = F.response) == null ? void 0 : A.data).message);
18347
+ a.setError(((w = F.response) == null ? void 0 : w.data).message);
18347
18348
  } else
18348
18349
  a.setError("There was an issue, please contact helpmate");
18349
18350
  }
18350
18351
  }, C = async () => {
18351
- var y, P, w, I, A, O, F, Q;
18352
+ var y, P, A, I, w, O, F, Q;
18352
18353
  try {
18353
18354
  const G = await Ce({
18354
18355
  url: "https://api-v2-dev.genpact.com/api/delphi/v2/myprofile",
@@ -18358,7 +18359,7 @@ const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t)
18358
18359
  },
18359
18360
  method: "get"
18360
18361
  });
18361
- c.genomeFunctionValue = (w = (P = (y = G == null ? void 0 : G.data) == null ? void 0 : y.role_function) == null ? void 0 : P.function) == null ? void 0 : w.name, c.genomeRoleValue = (O = (A = (I = G == null ? void 0 : G.data) == null ? void 0 : I.role_function) == null ? void 0 : A.role) == null ? void 0 : O.name, console.log("profile data", c.genomeFunctionValue, c.genomeRoleValue);
18362
+ c.genomeFunctionValue = (A = (P = (y = G == null ? void 0 : G.data) == null ? void 0 : y.role_function) == null ? void 0 : P.function) == null ? void 0 : A.name, c.genomeRoleValue = (O = (w = (I = G == null ? void 0 : G.data) == null ? void 0 : I.role_function) == null ? void 0 : w.role) == null ? void 0 : O.name, console.log("profile data", c.genomeFunctionValue, c.genomeRoleValue);
18362
18363
  } catch (G) {
18363
18364
  if (console.error(G), e.value = !1, Ce.isAxiosError(G)) {
18364
18365
  const K = G;
@@ -18370,28 +18371,28 @@ const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t)
18370
18371
  return (y, P) => (N(), j("section", null, [
18371
18372
  ge(sp, { "is-visible": i.value }, {
18372
18373
  default: op(() => [
18373
- xy
18374
+ qy
18374
18375
  ]),
18375
18376
  _: 1
18376
18377
  }, 8, ["is-visible"]),
18377
18378
  R("div", null, [
18378
- e.value ? (N(), Al(yl, { key: 0 })) : (N(), j("main", qy, [
18379
- R("div", By, [
18379
+ e.value ? (N(), Al(yl, { key: 0 })) : (N(), j("main", By, [
18380
+ R("div", Fy, [
18380
18381
  ge(zl)
18381
18382
  ]),
18382
- Fy,
18383
- R("div", Uy, [
18384
- R("div", Hy, [
18385
- R("div", Wy, [
18383
+ Uy,
18384
+ R("div", Hy, [
18385
+ R("div", Wy, [
18386
+ R("div", Vy, [
18386
18387
  ge(_C),
18387
18388
  R("div", {
18388
18389
  class: kt(k(c).showVoiceComponent === !0 ? "mt-[160px]" : "mt-[120px]")
18389
18390
  }, null, 2)
18390
18391
  ]),
18391
- k(c).assessmentQuestion ? (N(), j("div", Vy, [
18392
+ k(c).assessmentQuestion ? (N(), j("div", $y, [
18392
18393
  ge(Cy)
18393
- ])) : (N(), j("div", $y, [
18394
- ge(Ly)
18394
+ ])) : (N(), j("div", Ky, [
18395
+ ge(xy)
18395
18396
  ]))
18396
18397
  ])
18397
18398
  ])
@@ -18400,7 +18401,7 @@ const Ly = /* @__PURE__ */ Ft(jy, [["__scopeId", "data-v-b13dca42"]]), Xd = (t)
18400
18401
  ]));
18401
18402
  }
18402
18403
  });
18403
- const PR = /* @__PURE__ */ Ft(Ky, [["__scopeId", "data-v-904e6953"]]);
18404
+ const RR = /* @__PURE__ */ Ft(Jy, [["__scopeId", "data-v-904e6953"]]);
18404
18405
  export {
18405
- PR as default
18406
+ RR as default
18406
18407
  };