rasa-pro 3.12.6.dev2__py3-none-any.whl → 3.12.7__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of rasa-pro might be problematic. Click here for more details.

Files changed (107) hide show
  1. rasa/__init__.py +0 -6
  2. rasa/cli/run.py +10 -6
  3. rasa/cli/utils.py +7 -0
  4. rasa/core/actions/action.py +0 -6
  5. rasa/core/channels/channel.py +93 -0
  6. rasa/core/channels/inspector/dist/assets/{arc-c7691751.js → arc-351bec79.js} +1 -1
  7. rasa/core/channels/inspector/dist/assets/{blockDiagram-38ab4fdb-ab99dff7.js → blockDiagram-38ab4fdb-2567f3e5.js} +1 -1
  8. rasa/core/channels/inspector/dist/assets/{c4Diagram-3d4e48cf-08c35a6b.js → c4Diagram-3d4e48cf-c94acad0.js} +1 -1
  9. rasa/core/channels/inspector/dist/assets/channel-96a753ef.js +1 -0
  10. rasa/core/channels/inspector/dist/assets/{classDiagram-70f12bd4-9e9c71c9.js → classDiagram-70f12bd4-a2c4f658.js} +1 -1
  11. rasa/core/channels/inspector/dist/assets/{classDiagram-v2-f2320105-15e7e2bf.js → classDiagram-v2-f2320105-4036ee82.js} +1 -1
  12. rasa/core/channels/inspector/dist/assets/clone-5bbb0c7d.js +1 -0
  13. rasa/core/channels/inspector/dist/assets/{createText-2e5e7dd3-9c105cb1.js → createText-2e5e7dd3-01f8708a.js} +1 -1
  14. rasa/core/channels/inspector/dist/assets/{edges-e0da2a9e-77e89e48.js → edges-e0da2a9e-17b4c582.js} +1 -1
  15. rasa/core/channels/inspector/dist/assets/{erDiagram-9861fffd-7a011646.js → erDiagram-9861fffd-5b382730.js} +1 -1
  16. rasa/core/channels/inspector/dist/assets/{flowDb-956e92f1-b6f105ac.js → flowDb-956e92f1-c9dd4758.js} +1 -1
  17. rasa/core/channels/inspector/dist/assets/{flowDiagram-66a62f08-ce4f18c2.js → flowDiagram-66a62f08-014c7159.js} +1 -1
  18. rasa/core/channels/inspector/dist/assets/flowDiagram-v2-96b9c2cf-72082386.js +1 -0
  19. rasa/core/channels/inspector/dist/assets/{flowchart-elk-definition-4a651766-cb5f6da4.js → flowchart-elk-definition-4a651766-bc13fd64.js} +1 -1
  20. rasa/core/channels/inspector/dist/assets/{ganttDiagram-c361ad54-e4d19e28.js → ganttDiagram-c361ad54-a3bc832f.js} +1 -1
  21. rasa/core/channels/inspector/dist/assets/{gitGraphDiagram-72cf32ee-727b1c33.js → gitGraphDiagram-72cf32ee-4f0983dd.js} +1 -1
  22. rasa/core/channels/inspector/dist/assets/{graph-6e2ab9a7.js → graph-0069f93e.js} +1 -1
  23. rasa/core/channels/inspector/dist/assets/{index-3862675e-84ec700f.js → index-3862675e-7ddaa093.js} +1 -1
  24. rasa/core/channels/inspector/dist/assets/{index-098a1a24.js → index-d77a19b4.js} +129 -116
  25. rasa/core/channels/inspector/dist/assets/{infoDiagram-f8f76790-78dda442.js → infoDiagram-f8f76790-c3e28742.js} +1 -1
  26. rasa/core/channels/inspector/dist/assets/{journeyDiagram-49397b02-f1cc6dd1.js → journeyDiagram-49397b02-6d36c64c.js} +1 -1
  27. rasa/core/channels/inspector/dist/assets/{layout-d98dcd0c.js → layout-3d27f9c0.js} +1 -1
  28. rasa/core/channels/inspector/dist/assets/{line-838e3d82.js → line-e8cb25c5.js} +1 -1
  29. rasa/core/channels/inspector/dist/assets/{linear-eae72406.js → linear-4321e9fa.js} +1 -1
  30. rasa/core/channels/inspector/dist/assets/{mindmap-definition-fc14e90a-c96fd84b.js → mindmap-definition-fc14e90a-47e328b2.js} +1 -1
  31. rasa/core/channels/inspector/dist/assets/{pieDiagram-8a3498a8-c936d4e2.js → pieDiagram-8a3498a8-647edbaf.js} +1 -1
  32. rasa/core/channels/inspector/dist/assets/{quadrantDiagram-120e2f19-b338eb8f.js → quadrantDiagram-120e2f19-0703ad7d.js} +1 -1
  33. rasa/core/channels/inspector/dist/assets/{requirementDiagram-deff3bca-c6b6c0d5.js → requirementDiagram-deff3bca-59312f87.js} +1 -1
  34. rasa/core/channels/inspector/dist/assets/{sankeyDiagram-04a897e0-b9372e19.js → sankeyDiagram-04a897e0-8e170e1c.js} +1 -1
  35. rasa/core/channels/inspector/dist/assets/{sequenceDiagram-704730f1-479e0a3f.js → sequenceDiagram-704730f1-b2b42696.js} +1 -1
  36. rasa/core/channels/inspector/dist/assets/{stateDiagram-587899a1-fd26eebc.js → stateDiagram-587899a1-239f7e55.js} +1 -1
  37. rasa/core/channels/inspector/dist/assets/{stateDiagram-v2-d93cdb3a-3233e0ae.js → stateDiagram-v2-d93cdb3a-9cb9c726.js} +1 -1
  38. rasa/core/channels/inspector/dist/assets/{styles-6aaf32cf-1fdd392b.js → styles-6aaf32cf-d15a0f74.js} +1 -1
  39. rasa/core/channels/inspector/dist/assets/{styles-9a916d00-6d7bfa1b.js → styles-9a916d00-d7c52634.js} +1 -1
  40. rasa/core/channels/inspector/dist/assets/{styles-c10674c1-f86aab11.js → styles-c10674c1-cf79ea88.js} +1 -1
  41. rasa/core/channels/inspector/dist/assets/{svgDrawCommon-08f97a94-e3e49d7a.js → svgDrawCommon-08f97a94-52b3b9f9.js} +1 -1
  42. rasa/core/channels/inspector/dist/assets/{timeline-definition-85554ec2-6fe08b4d.js → timeline-definition-85554ec2-c8e3cd8c.js} +1 -1
  43. rasa/core/channels/inspector/dist/assets/{xychartDiagram-e933f94c-c2e06fd6.js → xychartDiagram-e933f94c-300afa53.js} +1 -1
  44. rasa/core/channels/inspector/dist/index.html +1 -1
  45. rasa/core/channels/inspector/src/components/Chat.tsx +23 -2
  46. rasa/core/channels/inspector/src/components/DiagramFlow.tsx +2 -5
  47. rasa/core/channels/inspector/src/helpers/conversation.ts +16 -0
  48. rasa/core/channels/inspector/src/types.ts +1 -1
  49. rasa/core/channels/voice_ready/audiocodes.py +53 -21
  50. rasa/core/channels/voice_ready/jambonz.py +25 -5
  51. rasa/core/channels/voice_ready/jambonz_protocol.py +4 -0
  52. rasa/core/channels/voice_ready/twilio_voice.py +48 -1
  53. rasa/core/channels/voice_stream/tts/azure.py +11 -2
  54. rasa/core/channels/voice_stream/twilio_media_streams.py +101 -26
  55. rasa/core/nlg/contextual_response_rephraser.py +4 -21
  56. rasa/core/nlg/summarize.py +1 -15
  57. rasa/core/policies/enterprise_search_policy.py +3 -16
  58. rasa/core/policies/flows/flow_executor.py +3 -38
  59. rasa/core/policies/intentless_policy.py +4 -17
  60. rasa/core/policies/policy.py +0 -2
  61. rasa/core/processor.py +19 -5
  62. rasa/core/utils.py +53 -0
  63. rasa/dialogue_understanding/coexistence/llm_based_router.py +4 -18
  64. rasa/dialogue_understanding/commands/cancel_flow_command.py +4 -59
  65. rasa/dialogue_understanding/commands/start_flow_command.py +0 -41
  66. rasa/dialogue_understanding/generator/command_generator.py +67 -0
  67. rasa/dialogue_understanding/generator/llm_based_command_generator.py +4 -20
  68. rasa/dialogue_understanding/generator/llm_command_generator.py +1 -3
  69. rasa/dialogue_understanding/generator/single_step/compact_llm_command_generator.py +1 -12
  70. rasa/dialogue_understanding/patterns/default_flows_for_patterns.yml +0 -61
  71. rasa/dialogue_understanding/processor/command_processor.py +7 -65
  72. rasa/dialogue_understanding/stack/utils.py +0 -38
  73. rasa/e2e_test/utils/validation.py +3 -3
  74. rasa/hooks.py +0 -55
  75. rasa/shared/constants.py +0 -5
  76. rasa/shared/core/constants.py +0 -8
  77. rasa/shared/core/domain.py +12 -3
  78. rasa/shared/core/flows/flow.py +0 -17
  79. rasa/shared/core/flows/flows_yaml_schema.json +3 -38
  80. rasa/shared/core/flows/steps/collect.py +5 -18
  81. rasa/shared/core/flows/utils.py +1 -16
  82. rasa/shared/core/slot_mappings.py +11 -5
  83. rasa/shared/nlu/constants.py +0 -1
  84. rasa/shared/providers/constants.py +0 -9
  85. rasa/shared/providers/llm/_base_litellm_client.py +4 -14
  86. rasa/shared/providers/llm/litellm_router_llm_client.py +7 -17
  87. rasa/shared/providers/llm/llm_client.py +15 -24
  88. rasa/shared/providers/llm/self_hosted_llm_client.py +2 -10
  89. rasa/shared/utils/common.py +11 -1
  90. rasa/shared/utils/health_check/health_check.py +1 -7
  91. rasa/tracing/instrumentation/attribute_extractors.py +4 -4
  92. rasa/tracing/instrumentation/intentless_policy_instrumentation.py +1 -2
  93. rasa/utils/licensing.py +0 -15
  94. rasa/validator.py +1 -123
  95. rasa/version.py +1 -1
  96. {rasa_pro-3.12.6.dev2.dist-info → rasa_pro-3.12.7.dist-info}/METADATA +3 -4
  97. {rasa_pro-3.12.6.dev2.dist-info → rasa_pro-3.12.7.dist-info}/RECORD +100 -103
  98. rasa/core/actions/action_handle_digressions.py +0 -164
  99. rasa/core/channels/inspector/dist/assets/channel-11268142.js +0 -1
  100. rasa/core/channels/inspector/dist/assets/clone-ff7f2ce7.js +0 -1
  101. rasa/core/channels/inspector/dist/assets/flowDiagram-v2-96b9c2cf-cba7ae20.js +0 -1
  102. rasa/dialogue_understanding/commands/handle_digressions_command.py +0 -144
  103. rasa/dialogue_understanding/patterns/handle_digressions.py +0 -81
  104. rasa/monkey_patches.py +0 -91
  105. {rasa_pro-3.12.6.dev2.dist-info → rasa_pro-3.12.7.dist-info}/NOTICE +0 -0
  106. {rasa_pro-3.12.6.dev2.dist-info → rasa_pro-3.12.7.dist-info}/WHEEL +0 -0
  107. {rasa_pro-3.12.6.dev2.dist-info → rasa_pro-3.12.7.dist-info}/entry_points.txt +0 -0
@@ -4,12 +4,6 @@ import structlog
4
4
  from jinja2 import Template
5
5
 
6
6
  from rasa.core.tracker_store import DialogueStateTracker
7
- from rasa.shared.constants import (
8
- LANGFUSE_CUSTOM_METADATA_DICT,
9
- LANGFUSE_METADATA_SESSION_ID,
10
- LANGFUSE_METADATA_USER_ID,
11
- LANGFUSE_TAGS,
12
- )
13
7
  from rasa.shared.providers.llm.llm_client import LLMClient
14
8
  from rasa.shared.utils.llm import (
15
9
  tracker_as_readable_transcript,
@@ -52,8 +46,6 @@ async def summarize_conversation(
52
46
  tracker: DialogueStateTracker,
53
47
  llm: LLMClient,
54
48
  max_turns: Optional[int] = MAX_TURNS_DEFAULT,
55
- user_id: Optional[str] = None,
56
- sender_id: Optional[str] = None,
57
49
  ) -> str:
58
50
  """Summarizes the dialogue using the LLM.
59
51
 
@@ -66,14 +58,8 @@ async def summarize_conversation(
66
58
  The summary of the dialogue.
67
59
  """
68
60
  prompt = _create_summarization_prompt(tracker, max_turns)
69
- metadata = {
70
- LANGFUSE_METADATA_USER_ID: user_id or "unknown",
71
- LANGFUSE_METADATA_SESSION_ID: sender_id or "",
72
- LANGFUSE_CUSTOM_METADATA_DICT: {"component": "summarize_conversation"},
73
- LANGFUSE_TAGS: ["summarize_conversation"],
74
- }
75
61
  try:
76
- llm_response = await llm.acompletion(prompt, metadata)
62
+ llm_response = await llm.acompletion(prompt)
77
63
  summarization = llm_response.choices[0].strip()
78
64
  structlogger.debug(
79
65
  "summarization.success", summarization=summarization, prompt=prompt
@@ -46,10 +46,6 @@ from rasa.graph_components.providers.forms_provider import Forms
46
46
  from rasa.graph_components.providers.responses_provider import Responses
47
47
  from rasa.shared.constants import (
48
48
  EMBEDDINGS_CONFIG_KEY,
49
- LANGFUSE_CUSTOM_METADATA_DICT,
50
- LANGFUSE_METADATA_SESSION_ID,
51
- LANGFUSE_METADATA_USER_ID,
52
- LANGFUSE_TAGS,
53
49
  MODEL_CONFIG_KEY,
54
50
  MODEL_GROUP_ID_CONFIG_KEY,
55
51
  MODEL_NAME_CONFIG_KEY,
@@ -549,9 +545,7 @@ class EnterpriseSearchPolicy(LLMHealthCheckMixin, EmbeddingsHealthCheckMixin, Po
549
545
 
550
546
  if self.use_llm:
551
547
  prompt = self._render_prompt(tracker, documents.results)
552
- llm_response = await self._generate_llm_answer(
553
- llm, prompt, tracker.sender_id
554
- )
548
+ llm_response = await self._generate_llm_answer(llm, prompt)
555
549
  llm_response = LLMResponse.ensure_llm_response(llm_response)
556
550
 
557
551
  self._add_prompt_and_llm_response_to_latest_message(
@@ -647,26 +641,19 @@ class EnterpriseSearchPolicy(LLMHealthCheckMixin, EmbeddingsHealthCheckMixin, Po
647
641
 
648
642
  @measure_llm_latency
649
643
  async def _generate_llm_answer(
650
- self, llm: LLMClient, prompt: Text, sender_id: str
644
+ self, llm: LLMClient, prompt: Text
651
645
  ) -> Optional[LLMResponse]:
652
646
  """Fetches an LLM completion for the provided prompt.
653
647
 
654
648
  Args:
655
649
  llm: The LLM client used to get the completion.
656
650
  prompt: The prompt text to send to the model.
657
- sender_id: sender_id from the tracker.
658
651
 
659
652
  Returns:
660
653
  An LLMResponse object, or None if the call fails.
661
654
  """
662
- metadata = {
663
- LANGFUSE_METADATA_USER_ID: self.user_id,
664
- LANGFUSE_METADATA_SESSION_ID: sender_id,
665
- LANGFUSE_CUSTOM_METADATA_DICT: {"component": self.__class__.__name__},
666
- LANGFUSE_TAGS: [self.__class__.__name__],
667
- }
668
655
  try:
669
- return await llm.acompletion(prompt, metadata)
656
+ return await llm.acompletion(prompt)
670
657
  except Exception as e:
671
658
  # unfortunately, langchain does not wrap LLM exceptions which means
672
659
  # we have to catch all exceptions here
@@ -23,7 +23,6 @@ from rasa.core.policies.flows.flow_step_result import (
23
23
  )
24
24
  from rasa.dialogue_understanding.commands import CancelFlowCommand
25
25
  from rasa.dialogue_understanding.patterns.cancel import CancelPatternFlowStackFrame
26
- from rasa.dialogue_understanding.patterns.clarify import ClarifyPatternFlowStackFrame
27
26
  from rasa.dialogue_understanding.patterns.collect_information import (
28
27
  CollectInformationPatternFlowStackFrame,
29
28
  )
@@ -51,7 +50,6 @@ from rasa.dialogue_understanding.stack.frames.flow_stack_frame import (
51
50
  )
52
51
  from rasa.dialogue_understanding.stack.utils import (
53
52
  top_user_flow_frame,
54
- user_flows_on_the_stack,
55
53
  )
56
54
  from rasa.shared.constants import RASA_PATTERN_HUMAN_HANDOFF
57
55
  from rasa.shared.core.constants import (
@@ -280,33 +278,6 @@ def trigger_pattern_continue_interrupted(
280
278
  return events
281
279
 
282
280
 
283
- def trigger_pattern_clarification(
284
- current_frame: DialogueStackFrame, stack: DialogueStack, flows: FlowsList
285
- ) -> None:
286
- """Trigger the pattern to clarify which topic to continue if needed."""
287
- if not isinstance(current_frame, UserFlowStackFrame):
288
- return None
289
-
290
- if current_frame.frame_type in [
291
- FlowStackFrameType.CALL,
292
- FlowStackFrameType.INTERRUPT,
293
- ]:
294
- # we want to return to the flow that called
295
- # the current flow or the flow that was interrupted
296
- # by the current flow
297
- return None
298
-
299
- pending_flows = [
300
- flows.flow_by_id(frame.flow_id)
301
- for frame in stack.frames
302
- if isinstance(frame, UserFlowStackFrame)
303
- and frame.flow_id != current_frame.flow_id
304
- ]
305
-
306
- flow_names = [flow.readable_name() for flow in pending_flows if flow is not None]
307
- stack.push(ClarifyPatternFlowStackFrame(names=flow_names))
308
-
309
-
310
281
  def trigger_pattern_completed(
311
282
  current_frame: DialogueStackFrame, stack: DialogueStack, flows: FlowsList
312
283
  ) -> None:
@@ -675,15 +646,9 @@ def _run_end_step(
675
646
  structlogger.debug("flow.step.run.flow_end")
676
647
  current_frame = stack.pop()
677
648
  trigger_pattern_completed(current_frame, stack, flows)
678
- resumed_events = []
679
- if len(user_flows_on_the_stack(stack)) > 1:
680
- # if there are more user flows on the stack,
681
- # we need to trigger the pattern clarify
682
- trigger_pattern_clarification(current_frame, stack, flows)
683
- else:
684
- resumed_events = trigger_pattern_continue_interrupted(
685
- current_frame, stack, flows, tracker
686
- )
649
+ resumed_events = trigger_pattern_continue_interrupted(
650
+ current_frame, stack, flows, tracker
651
+ )
687
652
  reset_events: List[Event] = reset_scoped_slots(current_frame, flow, tracker)
688
653
  return ContinueFlowWithNextStep(
689
654
  events=initial_events + reset_events + resumed_events, has_flow_ended=True
@@ -30,10 +30,6 @@ from rasa.graph_components.providers.forms_provider import Forms
30
30
  from rasa.graph_components.providers.responses_provider import Responses
31
31
  from rasa.shared.constants import (
32
32
  EMBEDDINGS_CONFIG_KEY,
33
- LANGFUSE_CUSTOM_METADATA_DICT,
34
- LANGFUSE_METADATA_SESSION_ID,
35
- LANGFUSE_METADATA_USER_ID,
36
- LANGFUSE_TAGS,
37
33
  LLM_CONFIG_KEY,
38
34
  MODEL_CONFIG_KEY,
39
35
  MODEL_GROUP_ID_CONFIG_KEY,
@@ -623,7 +619,6 @@ class IntentlessPolicy(LLMHealthCheckMixin, EmbeddingsHealthCheckMixin, Policy):
623
619
  response_examples: List[str],
624
620
  conversation_samples: List[str],
625
621
  history: str,
626
- sender_id: str,
627
622
  ) -> Optional[str]:
628
623
  """Make the llm call to generate an answer."""
629
624
  llm = llm_factory(self.config.get(LLM_CONFIG_KEY), DEFAULT_LLM_CONFIG)
@@ -639,19 +634,11 @@ class IntentlessPolicy(LLMHealthCheckMixin, EmbeddingsHealthCheckMixin, Policy):
639
634
  log_event="intentless_policy.generate_answer.prompt_rendered",
640
635
  prompt=prompt,
641
636
  )
642
- return await self._generate_llm_answer(llm, prompt, sender_id)
637
+ return await self._generate_llm_answer(llm, prompt)
643
638
 
644
- async def _generate_llm_answer(
645
- self, llm: LLMClient, prompt: str, sender_id: str
646
- ) -> Optional[str]:
647
- metadata = {
648
- LANGFUSE_METADATA_USER_ID: self.user_id,
649
- LANGFUSE_METADATA_SESSION_ID: sender_id,
650
- LANGFUSE_CUSTOM_METADATA_DICT: {"component": self.__class__.__name__},
651
- LANGFUSE_TAGS: [self.__class__.__name__],
652
- }
639
+ async def _generate_llm_answer(self, llm: LLMClient, prompt: str) -> Optional[str]:
653
640
  try:
654
- llm_response = await llm.acompletion(prompt, metadata)
641
+ llm_response = await llm.acompletion(prompt)
655
642
  return llm_response.choices[0]
656
643
  except Exception as e:
657
644
  # unfortunately, langchain does not wrap LLM exceptions which means
@@ -727,7 +714,7 @@ class IntentlessPolicy(LLMHealthCheckMixin, EmbeddingsHealthCheckMixin, Policy):
727
714
  final_response_examples.append(resp)
728
715
 
729
716
  llm_response = await self.generate_answer(
730
- final_response_examples, conversation_samples, history, tracker.sender_id
717
+ final_response_examples, conversation_samples, history
731
718
  )
732
719
  if not llm_response:
733
720
  structlogger.debug("intentless_policy.prediction.skip_llm_fail")
@@ -39,7 +39,6 @@ from rasa.shared.core.generator import TrackerWithCachedStates
39
39
  from rasa.shared.core.trackers import DialogueStateTracker
40
40
  from rasa.shared.exceptions import FileIOException, RasaException
41
41
  from rasa.shared.nlu.constants import ACTION_NAME, ACTION_TEXT, ENTITIES, INTENT, TEXT
42
- from rasa.utils.licensing import get_human_readable_licence_owner
43
42
 
44
43
  if TYPE_CHECKING:
45
44
  from rasa.core.featurizers.tracker_featurizers import (
@@ -173,7 +172,6 @@ class Policy(GraphComponent):
173
172
 
174
173
  self._model_storage = model_storage
175
174
  self._resource = resource
176
- self.user_id = get_human_readable_licence_owner()
177
175
 
178
176
  @classmethod
179
177
  def create(
rasa/core/processor.py CHANGED
@@ -76,6 +76,7 @@ from rasa.shared.core.constants import (
76
76
  SLOT_SILENCE_TIMEOUT,
77
77
  USER_INTENT_RESTART,
78
78
  USER_INTENT_SILENCE_TIMEOUT,
79
+ SetSlotExtractor,
79
80
  )
80
81
  from rasa.shared.core.events import (
81
82
  ActionExecuted,
@@ -766,13 +767,26 @@ class MessageProcessor:
766
767
  if self.http_interpreter:
767
768
  parse_data = await self.http_interpreter.parse(message)
768
769
  else:
769
- regex_reader = create_regex_pattern_reader(message, self.domain)
770
-
771
770
  processed_message = Message({TEXT: message.text})
772
- if regex_reader:
773
- processed_message = regex_reader.unpack_regex_message(
774
- message=processed_message, domain=self.domain
771
+
772
+ all_flows = await self.get_flows()
773
+ should_force_slot_command, slot_name = (
774
+ rasa.core.utils.should_force_slot_filling(tracker, all_flows)
775
+ )
776
+
777
+ if should_force_slot_command:
778
+ command = SetSlotCommand(
779
+ name=slot_name,
780
+ value=message.text,
781
+ extractor=SetSlotExtractor.COMMAND_PAYLOAD_READER.value,
775
782
  )
783
+ processed_message.set(COMMANDS, [command.as_dict()], add_to_output=True)
784
+ else:
785
+ regex_reader = create_regex_pattern_reader(message, self.domain)
786
+ if regex_reader:
787
+ processed_message = regex_reader.unpack_regex_message(
788
+ message=processed_message, domain=self.domain
789
+ )
776
790
 
777
791
  # Invalid use of slash syntax, sanitize the message before passing
778
792
  # it to the graph
rasa/core/utils.py CHANGED
@@ -19,6 +19,7 @@ from rasa.core.constants import (
19
19
  )
20
20
  from rasa.core.lock_store import InMemoryLockStore, LockStore, RedisLockStore
21
21
  from rasa.shared.constants import DEFAULT_ENDPOINTS_PATH, TCP_PROTOCOL
22
+ from rasa.shared.core.constants import SlotMappingType
22
23
  from rasa.shared.core.trackers import DialogueStateTracker
23
24
  from rasa.utils.endpoints import (
24
25
  EndpointConfig,
@@ -30,6 +31,7 @@ from rasa.utils.io import write_yaml
30
31
  if TYPE_CHECKING:
31
32
  from rasa.core.nlg import NaturalLanguageGenerator
32
33
  from rasa.shared.core.domain import Domain
34
+ from rasa.shared.core.flows.flows_list import FlowsList
33
35
 
34
36
  structlogger = structlog.get_logger()
35
37
 
@@ -364,3 +366,54 @@ def add_bot_utterance_metadata(
364
366
  ]
365
367
 
366
368
  return message
369
+
370
+
371
+ def should_force_slot_filling(
372
+ tracker: Optional[DialogueStateTracker], flows: "FlowsList"
373
+ ) -> Tuple[bool, Optional[str]]:
374
+ """Check if the flow should force slot filling.
375
+
376
+ This is only valid when the flow is at a collect information step which
377
+ has set `force_slot_filling` to true and the slot has a valid `from_text` mapping.
378
+
379
+ Args:
380
+ tracker: The dialogue state tracker.
381
+ flows: The list of flows.
382
+
383
+ Returns:
384
+ A tuple of a boolean indicating if the flow should force slot filling
385
+ and the name of the slot if applicable.
386
+ """
387
+ from rasa.dialogue_understanding.processor.command_processor import (
388
+ get_current_collect_step,
389
+ )
390
+
391
+ if tracker is None:
392
+ structlogger.error(
393
+ "slot.force_slot_filling.error",
394
+ event_info="Tracker is None. Cannot force slot filling.",
395
+ )
396
+ return False, None
397
+
398
+ stack = tracker.stack
399
+ step = get_current_collect_step(stack, flows)
400
+ if step is None or not step.force_slot_filling:
401
+ return False, None
402
+
403
+ slot_name = step.collect
404
+ slot = tracker.slots.get(slot_name)
405
+
406
+ if not slot:
407
+ structlogger.debug(
408
+ "slot.force_slot_filling.error",
409
+ event_info=f"Slot '{slot_name}' not found in tracker. "
410
+ f"Cannot force slot filling. "
411
+ f"Please check if the slot is defined in the domain.",
412
+ )
413
+ return False, None
414
+
415
+ for slot_mapping in slot.mappings:
416
+ if slot_mapping.type == SlotMappingType.FROM_TEXT:
417
+ return True, slot_name
418
+
419
+ return False, None
@@ -23,10 +23,6 @@ from rasa.engine.recipes.default_recipe import DefaultV1Recipe
23
23
  from rasa.engine.storage.resource import Resource
24
24
  from rasa.engine.storage.storage import ModelStorage
25
25
  from rasa.shared.constants import (
26
- LANGFUSE_CUSTOM_METADATA_DICT,
27
- LANGFUSE_METADATA_SESSION_ID,
28
- LANGFUSE_METADATA_USER_ID,
29
- LANGFUSE_TAGS,
30
26
  MODEL_CONFIG_KEY,
31
27
  OPENAI_PROVIDER,
32
28
  PROMPT_CONFIG_KEY,
@@ -47,7 +43,6 @@ from rasa.shared.utils.llm import (
47
43
  llm_factory,
48
44
  resolve_model_client_config,
49
45
  )
50
- from rasa.utils.licensing import get_human_readable_licence_owner
51
46
  from rasa.utils.log_utils import log_llm
52
47
 
53
48
  LLM_BASED_ROUTER_PROMPT_FILE_NAME = "llm_based_router_prompt.jinja2"
@@ -118,7 +113,6 @@ class LLMBasedRouter(LLMHealthCheckMixin, GraphComponent):
118
113
  self._model_storage = model_storage
119
114
  self._resource = resource
120
115
  self.validate_config()
121
- self.user_id = get_human_readable_licence_owner()
122
116
 
123
117
  def validate_config(self) -> None:
124
118
  """Validate the config of the router."""
@@ -166,6 +160,7 @@ class LLMBasedRouter(LLMHealthCheckMixin, GraphComponent):
166
160
  **kwargs: Any,
167
161
  ) -> "LLMBasedRouter":
168
162
  """Loads trained component (see parent class for full docstring)."""
163
+
169
164
  # Perform health check on the resolved LLM client config
170
165
  llm_config = resolve_model_client_config(config.get(LLM_CONFIG_KEY, {}))
171
166
  cls.perform_llm_health_check(
@@ -237,7 +232,7 @@ class LLMBasedRouter(LLMHealthCheckMixin, GraphComponent):
237
232
  prompt=prompt,
238
233
  )
239
234
  # generating answer
240
- answer = await self._generate_answer_using_llm(prompt, tracker.sender_id)
235
+ answer = await self._generate_answer_using_llm(prompt)
241
236
  log_llm(
242
237
  logger=structlogger,
243
238
  log_module="LLMBasedRouter",
@@ -297,9 +292,7 @@ class LLMBasedRouter(LLMHealthCheckMixin, GraphComponent):
297
292
 
298
293
  return Template(self.prompt_template).render(**inputs)
299
294
 
300
- async def _generate_answer_using_llm(
301
- self, prompt: str, sender_id: str
302
- ) -> Optional[str]:
295
+ async def _generate_answer_using_llm(self, prompt: str) -> Optional[str]:
303
296
  """Use LLM to generate a response.
304
297
 
305
298
  Args:
@@ -310,15 +303,8 @@ class LLMBasedRouter(LLMHealthCheckMixin, GraphComponent):
310
303
  """
311
304
  llm = llm_factory(self.config.get(LLM_CONFIG_KEY), DEFAULT_LLM_CONFIG)
312
305
 
313
- metadata = {
314
- LANGFUSE_METADATA_USER_ID: self.user_id,
315
- LANGFUSE_METADATA_SESSION_ID: sender_id,
316
- LANGFUSE_CUSTOM_METADATA_DICT: {"component": self.__class__.__name__},
317
- LANGFUSE_TAGS: [self.__class__.__name__],
318
- }
319
-
320
306
  try:
321
- llm_response = await llm.acompletion(prompt, metadata)
307
+ llm_response = await llm.acompletion(prompt)
322
308
  return llm_response.choices[0]
323
309
  except Exception as e:
324
310
  # unfortunately, langchain does not wrap LLM exceptions which means
@@ -1,6 +1,5 @@
1
1
  from __future__ import annotations
2
2
 
3
- import copy
4
3
  import re
5
4
  from dataclasses import dataclass
6
5
  from typing import Any, Dict, List
@@ -13,10 +12,11 @@ from rasa.dialogue_understanding.commands.command_syntax_manager import (
13
12
  CommandSyntaxVersion,
14
13
  )
15
14
  from rasa.dialogue_understanding.patterns.cancel import CancelPatternFlowStackFrame
16
- from rasa.dialogue_understanding.patterns.clarify import ClarifyPatternFlowStackFrame
17
15
  from rasa.dialogue_understanding.stack.dialogue_stack import DialogueStack
18
- from rasa.dialogue_understanding.stack.frames import UserFlowStackFrame
19
- from rasa.dialogue_understanding.stack.frames.flow_stack_frame import FlowStackFrameType
16
+ from rasa.dialogue_understanding.stack.frames.flow_stack_frame import (
17
+ FlowStackFrameType,
18
+ UserFlowStackFrame,
19
+ )
20
20
  from rasa.dialogue_understanding.stack.utils import top_user_flow_frame
21
21
  from rasa.shared.core.events import Event, FlowCancelled
22
22
  from rasa.shared.core.flows import FlowsList
@@ -95,8 +95,6 @@ class CancelFlowCommand(Command):
95
95
  original_stack = original_tracker.stack
96
96
 
97
97
  applied_events: List[Event] = []
98
- # capture the top frame before we push new frames onto the stack
99
- initial_top_frame = stack.top()
100
98
  user_frame = top_user_flow_frame(original_stack)
101
99
  current_flow = user_frame.flow(all_flows) if user_frame else None
102
100
 
@@ -123,21 +121,6 @@ class CancelFlowCommand(Command):
123
121
  if user_frame:
124
122
  applied_events.append(FlowCancelled(user_frame.flow_id, user_frame.step_id))
125
123
 
126
- if initial_top_frame and isinstance(
127
- initial_top_frame, ClarifyPatternFlowStackFrame
128
- ):
129
- structlogger.debug(
130
- "command_executor.cancel_flow.cancel_clarification_options",
131
- clarification_options=initial_top_frame.clarification_options,
132
- )
133
- applied_events += cancel_all_pending_clarification_options(
134
- initial_top_frame,
135
- original_stack,
136
- canceled_frames,
137
- all_flows,
138
- stack,
139
- )
140
-
141
124
  return applied_events + tracker.create_stack_updated_events(stack)
142
125
 
143
126
  def __hash__(self) -> int:
@@ -172,41 +155,3 @@ class CancelFlowCommand(Command):
172
155
  CommandSyntaxManager.get_syntax_version(),
173
156
  mapper[CommandSyntaxManager.get_default_syntax_version()],
174
157
  )
175
-
176
-
177
- def cancel_all_pending_clarification_options(
178
- initial_top_frame: ClarifyPatternFlowStackFrame,
179
- original_stack: DialogueStack,
180
- canceled_frames: List[str],
181
- all_flows: FlowsList,
182
- stack: DialogueStack,
183
- ) -> List[FlowCancelled]:
184
- """Cancel all pending clarification options.
185
-
186
- This is a special case when the assistant asks the user to clarify
187
- which pending digression flow to start after the completion of an active flow.
188
- If the user chooses to cancel all options, this function takes care of
189
- updating the stack by removing all pending flow stack frames
190
- listed as clarification options.
191
- """
192
- clarification_names = set(initial_top_frame.names)
193
- to_be_canceled_frames = []
194
- applied_events = []
195
- for frame in reversed(original_stack.frames):
196
- if frame.frame_id in canceled_frames:
197
- continue
198
-
199
- to_be_canceled_frames.append(frame.frame_id)
200
- if isinstance(frame, UserFlowStackFrame):
201
- readable_flow_name = frame.flow(all_flows).readable_name()
202
- if readable_flow_name in clarification_names:
203
- stack.push(
204
- CancelPatternFlowStackFrame(
205
- canceled_name=readable_flow_name,
206
- canceled_frames=copy.deepcopy(to_be_canceled_frames),
207
- )
208
- )
209
- applied_events.append(FlowCancelled(frame.flow_id, frame.step_id))
210
- to_be_canceled_frames.clear()
211
-
212
- return applied_events
@@ -11,11 +11,6 @@ from rasa.dialogue_understanding.commands.command_syntax_manager import (
11
11
  CommandSyntaxManager,
12
12
  CommandSyntaxVersion,
13
13
  )
14
- from rasa.dialogue_understanding.patterns.clarify import FLOW_PATTERN_CLARIFICATION
15
- from rasa.dialogue_understanding.patterns.continue_interrupted import (
16
- ContinueInterruptedPatternFlowStackFrame,
17
- )
18
- from rasa.dialogue_understanding.stack.dialogue_stack import DialogueStack
19
14
  from rasa.dialogue_understanding.stack.frames.flow_stack_frame import (
20
15
  FlowStackFrameType,
21
16
  UserFlowStackFrame,
@@ -77,10 +72,6 @@ class StartFlowCommand(Command):
77
72
  applied_events: List[Event] = []
78
73
 
79
74
  if self.flow in user_flows_on_the_stack(stack):
80
- top_frame = stack.top()
81
- if top_frame is not None and top_frame.type() == FLOW_PATTERN_CLARIFICATION:
82
- return self.change_flow_frame_position_in_the_stack(stack, tracker)
83
-
84
75
  structlogger.debug(
85
76
  "command_executor.skip_command.already_started_flow", command=self
86
77
  )
@@ -149,35 +140,3 @@ class StartFlowCommand(Command):
149
140
  CommandSyntaxManager.get_syntax_version(),
150
141
  mapper[CommandSyntaxManager.get_default_syntax_version()],
151
142
  )
152
-
153
- def change_flow_frame_position_in_the_stack(
154
- self, stack: DialogueStack, tracker: DialogueStateTracker
155
- ) -> List[Event]:
156
- """Changes the position of the flow frame in the stack.
157
-
158
- This is a special case when pattern clarification is the active flow and
159
- the same flow is selected to start. In this case, the existing flow frame
160
- should be moved up in the stack.
161
- """
162
- frames = stack.frames[:]
163
-
164
- for idx, frame in enumerate(frames):
165
- if isinstance(frame, UserFlowStackFrame) and frame.flow_id == self.flow:
166
- structlogger.debug(
167
- "command_executor.change_flow_position_during_clarification",
168
- command=self,
169
- index=idx,
170
- )
171
- # pop the continue interrupted flow frame if it exists
172
- next_frame = frames[idx + 1] if idx + 1 < len(frames) else None
173
- if (
174
- isinstance(next_frame, ContinueInterruptedPatternFlowStackFrame)
175
- and next_frame.previous_flow_name == self.flow
176
- ):
177
- stack.frames.pop(idx + 1)
178
- # move up the existing flow from the stack
179
- stack.frames.pop(idx)
180
- stack.push(frame)
181
- return tracker.create_stack_updated_events(stack)
182
-
183
- return []
@@ -4,6 +4,7 @@ from typing import Any, Dict, List, Optional, Set, Text, Tuple
4
4
  import structlog
5
5
 
6
6
  from rasa.dialogue_understanding.commands import (
7
+ CannotHandleCommand,
7
8
  Command,
8
9
  CorrectSlotsCommand,
9
10
  ErrorCommand,
@@ -107,6 +108,14 @@ class CommandGenerator:
107
108
  commands = self._check_commands_against_startable_flows(
108
109
  commands, startable_flows
109
110
  )
111
+
112
+ # During force slot filling, keep only the command that sets the
113
+ # slot asked by the active collect step.
114
+ # Or return a CannotHandleCommand if no matching command is found.
115
+ commands = self._filter_commands_during_force_slot_filling(
116
+ commands, available_flows, tracker
117
+ )
118
+
110
119
  commands_dicts = [command.as_dict() for command in commands]
111
120
  message.set(COMMANDS, commands_dicts, add_to_output=True)
112
121
 
@@ -370,6 +379,64 @@ class CommandGenerator:
370
379
  Command.command_from_json(command) for command in message.get(COMMANDS, [])
371
380
  ]
372
381
 
382
+ @staticmethod
383
+ def _filter_commands_during_force_slot_filling(
384
+ commands: List[Command],
385
+ available_flows: FlowsList,
386
+ tracker: Optional[DialogueStateTracker] = None,
387
+ ) -> List[Command]:
388
+ """Filter commands during a collect step that has set `force_slot_filling`.
389
+
390
+ Args:
391
+ commands: The commands to filter.
392
+ available_flows: The available flows.
393
+ tracker: The tracker.
394
+
395
+ Returns:
396
+ The filtered commands.
397
+ """
398
+ from rasa.dialogue_understanding.processor.command_processor import (
399
+ get_current_collect_step,
400
+ )
401
+
402
+ if tracker is None:
403
+ structlogger.error(
404
+ "command_generator.filter_commands_during_force_slot_filling.tracker_not_found",
405
+ )
406
+ return commands
407
+
408
+ stack = tracker.stack
409
+ step = get_current_collect_step(stack, available_flows)
410
+
411
+ if step is None or not step.force_slot_filling:
412
+ return commands
413
+
414
+ # Retain only the command that sets the slot asked by
415
+ # the active collect step
416
+ filtered_commands: List[Command] = [
417
+ command
418
+ for command in commands
419
+ if (isinstance(command, SetSlotCommand) and command.name == step.collect)
420
+ ]
421
+
422
+ if not filtered_commands:
423
+ # If no commands were predicted, we need to return a CannotHandleCommand
424
+ structlogger.debug(
425
+ "command_generator.filter_commands_during_force_slot_filling.no_commands",
426
+ event_info=f"The command generator did not find any SetSlot "
427
+ f"command at the collect step for the slot '{step.collect}'. "
428
+ f"Returning a CannotHandleCommand instead.",
429
+ )
430
+ return [CannotHandleCommand()]
431
+
432
+ structlogger.debug(
433
+ "command_generator.filter_commands_during_force_slot_filling.filtered_commands",
434
+ slot_name=step.collect,
435
+ filtered_commands=filtered_commands,
436
+ )
437
+
438
+ return filtered_commands
439
+
373
440
 
374
441
  def gather_slot_names(commands: List[Command]) -> Set[str]:
375
442
  """Gather all slot names from the commands."""