datarobot-genai 0.1.75__py3-none-any.whl → 0.2.11__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (32) hide show
  1. datarobot_genai/core/agents/base.py +2 -1
  2. datarobot_genai/core/chat/responses.py +32 -4
  3. datarobot_genai/drmcp/core/config.py +52 -0
  4. datarobot_genai/drmcp/core/dr_mcp_server.py +45 -8
  5. datarobot_genai/drmcp/core/dynamic_prompts/dr_lib.py +22 -80
  6. datarobot_genai/drmcp/core/dynamic_prompts/register.py +4 -5
  7. datarobot_genai/drmcp/core/mcp_instance.py +41 -2
  8. datarobot_genai/drmcp/core/routes.py +4 -1
  9. datarobot_genai/drmcp/core/tool_config.py +95 -0
  10. datarobot_genai/drmcp/test_utils/mcp_utils_ete.py +29 -0
  11. datarobot_genai/drmcp/test_utils/openai_llm_mcp_client.py +6 -1
  12. datarobot_genai/drmcp/tools/clients/__init__.py +14 -0
  13. datarobot_genai/drmcp/tools/clients/atlassian.py +188 -0
  14. datarobot_genai/drmcp/tools/clients/confluence.py +196 -0
  15. datarobot_genai/drmcp/tools/clients/jira.py +147 -0
  16. datarobot_genai/drmcp/tools/clients/s3.py +28 -0
  17. datarobot_genai/drmcp/tools/confluence/__init__.py +14 -0
  18. datarobot_genai/drmcp/tools/confluence/tools.py +81 -0
  19. datarobot_genai/drmcp/tools/jira/__init__.py +14 -0
  20. datarobot_genai/drmcp/tools/jira/tools.py +52 -0
  21. datarobot_genai/drmcp/tools/predictive/predict.py +1 -1
  22. datarobot_genai/drmcp/tools/predictive/predict_realtime.py +1 -1
  23. datarobot_genai/langgraph/agent.py +143 -42
  24. datarobot_genai/nat/agent.py +4 -0
  25. datarobot_genai/nat/datarobot_auth_provider.py +110 -0
  26. datarobot_genai/nat/datarobot_mcp_client.py +234 -0
  27. {datarobot_genai-0.1.75.dist-info → datarobot_genai-0.2.11.dist-info}/METADATA +9 -2
  28. {datarobot_genai-0.1.75.dist-info → datarobot_genai-0.2.11.dist-info}/RECORD +32 -20
  29. {datarobot_genai-0.1.75.dist-info → datarobot_genai-0.2.11.dist-info}/entry_points.txt +2 -0
  30. {datarobot_genai-0.1.75.dist-info → datarobot_genai-0.2.11.dist-info}/WHEEL +0 -0
  31. {datarobot_genai-0.1.75.dist-info → datarobot_genai-0.2.11.dist-info}/licenses/AUTHORS +0 -0
  32. {datarobot_genai-0.1.75.dist-info → datarobot_genai-0.2.11.dist-info}/licenses/LICENSE +0 -0
@@ -23,6 +23,7 @@ from typing import TypedDict
23
23
  from typing import TypeVar
24
24
  from typing import cast
25
25
 
26
+ from ag_ui.core import Event
26
27
  from openai.types.chat import CompletionCreateParams
27
28
  from ragas import MultiTurnSample
28
29
 
@@ -167,7 +168,7 @@ class UsageMetrics(TypedDict):
167
168
 
168
169
  # Canonical return type for DRUM-compatible invoke implementations
169
170
  InvokeReturn = (
170
- AsyncGenerator[tuple[str, MultiTurnSample | None, UsageMetrics], None]
171
+ AsyncGenerator[tuple[str | Event, MultiTurnSample | None, UsageMetrics], None]
171
172
  | tuple[str, MultiTurnSample | None, UsageMetrics]
172
173
  )
173
174
 
@@ -27,6 +27,10 @@ from concurrent.futures import ThreadPoolExecutor
27
27
  from typing import Any
28
28
  from typing import TypeVar
29
29
 
30
+ from ag_ui.core import BaseEvent
31
+ from ag_ui.core import Event
32
+ from ag_ui.core import TextMessageChunkEvent
33
+ from ag_ui.core import TextMessageContentEvent
30
34
  from openai.types import CompletionUsage
31
35
  from openai.types.chat import ChatCompletion
32
36
  from openai.types.chat import ChatCompletionChunk
@@ -45,6 +49,7 @@ class CustomModelChatResponse(ChatCompletion):
45
49
 
46
50
  class CustomModelStreamingResponse(ChatCompletionChunk):
47
51
  pipeline_interactions: str | None = None
52
+ event: Event | None = None
48
53
 
49
54
 
50
55
  def to_custom_model_chat_response(
@@ -88,7 +93,7 @@ def to_custom_model_streaming_response(
88
93
  thread_pool_executor: ThreadPoolExecutor,
89
94
  event_loop: AbstractEventLoop,
90
95
  streaming_response_generator: AsyncGenerator[
91
- tuple[str, MultiTurnSample | None, dict[str, int]], None
96
+ tuple[str | Event, MultiTurnSample | None, dict[str, int]], None
92
97
  ],
93
98
  model: str | object | None,
94
99
  ) -> Iterator[CustomModelStreamingResponse]:
@@ -110,7 +115,7 @@ def to_custom_model_streaming_response(
110
115
  while True:
111
116
  try:
112
117
  (
113
- response_text,
118
+ response_text_or_event,
114
119
  pipeline_interactions,
115
120
  usage_metrics,
116
121
  ) = thread_pool_executor.submit(
@@ -119,10 +124,10 @@ def to_custom_model_streaming_response(
119
124
  last_pipeline_interactions = pipeline_interactions
120
125
  last_usage_metrics = usage_metrics
121
126
 
122
- if response_text:
127
+ if isinstance(response_text_or_event, str) and response_text_or_event:
123
128
  choice = ChunkChoice(
124
129
  index=0,
125
- delta=ChoiceDelta(role="assistant", content=response_text),
130
+ delta=ChoiceDelta(role="assistant", content=response_text_or_event),
126
131
  finish_reason=None,
127
132
  )
128
133
  yield CustomModelStreamingResponse(
@@ -135,6 +140,29 @@ def to_custom_model_streaming_response(
135
140
  if usage_metrics
136
141
  else None,
137
142
  )
143
+ elif isinstance(response_text_or_event, BaseEvent):
144
+ content = ""
145
+ if isinstance(
146
+ response_text_or_event, (TextMessageContentEvent, TextMessageChunkEvent)
147
+ ):
148
+ content = response_text_or_event.delta or content
149
+ choice = ChunkChoice(
150
+ index=0,
151
+ delta=ChoiceDelta(role="assistant", content=content),
152
+ finish_reason=None,
153
+ )
154
+
155
+ yield CustomModelStreamingResponse(
156
+ id=completion_id,
157
+ object="chat.completion.chunk",
158
+ created=created,
159
+ model=model,
160
+ choices=[choice],
161
+ usage=CompletionUsage.model_validate(required_usage_metrics | usage_metrics)
162
+ if usage_metrics
163
+ else None,
164
+ event=response_text_or_event,
165
+ )
138
166
  except StopAsyncIteration:
139
167
  break
140
168
  event_loop.run_until_complete(streaming_response_generator.aclose())
@@ -197,6 +197,54 @@ class MCPServerConfig(BaseSettings):
197
197
  description="Enable/disable predictive tools",
198
198
  )
199
199
 
200
+ # Jira tools
201
+ enable_jira_tools: bool = Field(
202
+ default=False,
203
+ validation_alias=AliasChoices(
204
+ RUNTIME_PARAM_ENV_VAR_NAME_PREFIX + "ENABLE_JIRA_TOOLS",
205
+ "ENABLE_JIRA_TOOLS",
206
+ ),
207
+ description="Enable/disable Jira tools",
208
+ )
209
+ is_jira_oauth_provider_configured: bool = Field(
210
+ default=False,
211
+ validation_alias=AliasChoices(
212
+ RUNTIME_PARAM_ENV_VAR_NAME_PREFIX + "IS_JIRA_OAUTH_PROVIDER_CONFIGURED",
213
+ "IS_JIRA_OAUTH_PROVIDER_CONFIGURED",
214
+ ),
215
+ description="Whether Jira OAuth provider is configured for Jira integration",
216
+ )
217
+
218
+ @property
219
+ def is_jira_oauth_configured(self) -> bool:
220
+ return self.is_jira_oauth_provider_configured or bool(
221
+ os.getenv("JIRA_CLIENT_ID") and os.getenv("JIRA_CLIENT_SECRET")
222
+ )
223
+
224
+ # Confluence tools
225
+ enable_confluence_tools: bool = Field(
226
+ default=False,
227
+ validation_alias=AliasChoices(
228
+ RUNTIME_PARAM_ENV_VAR_NAME_PREFIX + "ENABLE_CONFLUENCE_TOOLS",
229
+ "ENABLE_CONFLUENCE_TOOLS",
230
+ ),
231
+ description="Enable/disable Confluence tools",
232
+ )
233
+ is_confluence_oauth_provider_configured: bool = Field(
234
+ default=False,
235
+ validation_alias=AliasChoices(
236
+ RUNTIME_PARAM_ENV_VAR_NAME_PREFIX + "IS_CONFLUENCE_OAUTH_PROVIDER_CONFIGURED",
237
+ "IS_CONFLUENCE_OAUTH_PROVIDER_CONFIGURED",
238
+ ),
239
+ description="Whether Confluence OAuth provider is configured for Confluence integration",
240
+ )
241
+
242
+ @property
243
+ def is_confluence_oauth_configured(self) -> bool:
244
+ return self.is_confluence_oauth_provider_configured or bool(
245
+ os.getenv("CONFLUENCE_CLIENT_ID") and os.getenv("CONFLUENCE_CLIENT_SECRET")
246
+ )
247
+
200
248
  @field_validator(
201
249
  "otel_attributes",
202
250
  mode="before",
@@ -220,6 +268,10 @@ class MCPServerConfig(BaseSettings):
220
268
  "tool_registration_duplicate_behavior",
221
269
  "mcp_server_register_dynamic_prompts_on_startup",
222
270
  "enable_predictive_tools",
271
+ "enable_jira_tools",
272
+ "is_jira_oauth_provider_configured",
273
+ "enable_confluence_tools",
274
+ "is_confluence_oauth_provider_configured",
223
275
  mode="before",
224
276
  )
225
277
  @classmethod
@@ -40,6 +40,8 @@ from .routes_utils import prefix_mount_path
40
40
  from .server_life_cycle import BaseServerLifecycle
41
41
  from .telemetry import OtelASGIMiddleware
42
42
  from .telemetry import initialize_telemetry
43
+ from .tool_config import TOOL_CONFIGS
44
+ from .tool_config import is_tool_enabled
43
45
 
44
46
 
45
47
  def _import_modules_from_dir(
@@ -115,6 +117,9 @@ class DataRobotMCPServer:
115
117
  self._mcp = mcp
116
118
  self._mcp_transport = transport
117
119
 
120
+ # Configure MCP server capabilities
121
+ self._configure_mcp_capabilities()
122
+
118
123
  # Initialize telemetry
119
124
  initialize_telemetry(mcp)
120
125
 
@@ -139,11 +144,12 @@ class DataRobotMCPServer:
139
144
 
140
145
  # Load static tools modules
141
146
  base_dir = os.path.dirname(os.path.dirname(__file__))
142
- if self._config.enable_predictive_tools:
143
- _import_modules_from_dir(
144
- os.path.join(base_dir, "tools", "predictive"),
145
- "datarobot_genai.drmcp.tools.predictive",
146
- )
147
+ for tool_type, tool_config in TOOL_CONFIGS.items():
148
+ if is_tool_enabled(tool_type, self._config):
149
+ _import_modules_from_dir(
150
+ os.path.join(base_dir, "tools", tool_config["directory"]),
151
+ tool_config["package_prefix"],
152
+ )
147
153
 
148
154
  # Load memory management tools if available
149
155
  if self._memory_manager:
@@ -163,6 +169,37 @@ class DataRobotMCPServer:
163
169
  if transport == "streamable-http":
164
170
  register_routes(self._mcp)
165
171
 
172
+ def _configure_mcp_capabilities(self) -> None:
173
+ """Configure MCP capabilities that FastMCP doesn't expose directly.
174
+
175
+ See: https://github.com/modelcontextprotocol/python-sdk/issues/1126
176
+ """
177
+ server = self._mcp._mcp_server
178
+
179
+ # Declare prompts_changed capability (capabilities.prompts.listChanged: true)
180
+ server.notification_options.prompts_changed = True
181
+
182
+ # Declare experimental capabilities ( experimental.dynamic_prompts: true)
183
+ server.experimental_capabilities = {"dynamic_prompts": {"enabled": True}}
184
+
185
+ # Patch to include experimental_capabilities (FastMCP doesn't expose this)
186
+ original = server.create_initialization_options
187
+
188
+ def patched(
189
+ notification_options: Any = None,
190
+ experimental_capabilities: dict[str, dict[str, Any]] | None = None,
191
+ **kwargs: Any,
192
+ ) -> Any:
193
+ if experimental_capabilities is None:
194
+ experimental_capabilities = getattr(server, "experimental_capabilities", None)
195
+ return original(
196
+ notification_options=notification_options,
197
+ experimental_capabilities=experimental_capabilities,
198
+ **kwargs,
199
+ )
200
+
201
+ server.create_initialization_options = patched
202
+
166
203
  def run(self, show_banner: bool = False) -> None:
167
204
  """Run the DataRobot MCP server synchronously."""
168
205
  try:
@@ -179,6 +216,9 @@ class DataRobotMCPServer:
179
216
  self._logger.info("Registering dynamic prompts from prompt management...")
180
217
  asyncio.run(register_prompts_from_datarobot_prompt_management())
181
218
 
219
+ # Execute pre-server start actions
220
+ asyncio.run(self._lifecycle.pre_server_start(self._mcp))
221
+
182
222
  # List registered tools, prompts, and resources before starting server
183
223
  tools = asyncio.run(self._mcp._list_tools_mcp())
184
224
  prompts = asyncio.run(self._mcp._list_prompts_mcp())
@@ -198,9 +238,6 @@ class DataRobotMCPServer:
198
238
  for resource in resources:
199
239
  self._logger.info(f" > {resource.name}")
200
240
 
201
- # Execute pre-server start actions
202
- asyncio.run(self._lifecycle.pre_server_start(self._mcp))
203
-
204
241
  # Create event loop for async operations
205
242
  loop = asyncio.new_event_loop()
206
243
  asyncio.set_event_loop(loop)
@@ -12,78 +12,23 @@
12
12
  # See the License for the specific language governing permissions and
13
13
  # limitations under the License.
14
14
  from collections import defaultdict
15
- from dataclasses import dataclass
16
15
 
17
16
  import datarobot as dr
18
17
 
19
18
  from datarobot_genai.drmcp.core.clients import get_api_client
20
19
 
21
- # Needed SDK version (3.10.0) is not published yet. We'll reimplement simplified version of it.
22
- # get_datarobot_prompt_templates = dr.genai.PromptTemplate.list()
23
- # DrPrompt = dr.genai.PromptTemplate
24
- # DrPromptVersion = dr.genai.PromptTemplateVersion
25
- # DrVariable = dr.genai.Variable
26
20
 
27
-
28
- @dataclass
29
- class DrVariable:
30
- name: str
31
- description: str
32
-
33
-
34
- @dataclass
35
- class DrPromptVersion:
36
- id: str
37
- prompt_template_id: str
38
- version: int
39
- prompt_text: str
40
- variables: list[DrVariable]
41
-
42
- @classmethod
43
- def from_dict(cls, d: dict) -> "DrPromptVersion":
44
- variables = [
45
- DrVariable(name=v["name"], description=v["description"]) for v in d["variables"]
46
- ]
47
- return cls(
48
- id=d["id"],
49
- prompt_template_id=d["promptTemplateId"],
50
- version=d["version"],
51
- prompt_text=d["promptText"],
52
- variables=variables,
53
- )
54
-
55
-
56
- @dataclass
57
- class DrPrompt:
58
- id: str
59
- name: str
60
- description: str
61
-
62
- def get_latest_version(self) -> DrPromptVersion | None:
63
- all_prompt_template_versions = get_datarobot_prompt_template_versions([self.id])
64
- prompt_template_versions = all_prompt_template_versions.get(self.id)
65
-
66
- if not prompt_template_versions:
67
- return None
68
- latest_version = max(prompt_template_versions, key=lambda v: v.version)
69
- return latest_version
70
-
71
- @classmethod
72
- def from_dict(cls, d: dict) -> "DrPrompt":
73
- return cls(id=d["id"], name=d["name"], description=d["description"])
74
-
75
-
76
- def get_datarobot_prompt_templates() -> list[DrPrompt]:
77
- prompt_templates_data = dr.utils.pagination.unpaginate(
78
- initial_url="genai/promptTemplates/", initial_params={}, client=get_api_client()
79
- )
80
-
81
- return [DrPrompt.from_dict(prompt_template) for prompt_template in prompt_templates_data]
21
+ def get_datarobot_prompt_templates() -> list[dr.genai.PromptTemplate]:
22
+ try:
23
+ return dr.genai.PromptTemplate.list()
24
+ except Exception:
25
+ return []
82
26
 
83
27
 
84
28
  def get_datarobot_prompt_template_versions(
85
29
  prompt_template_ids: list[str],
86
- ) -> dict[str, list[DrPromptVersion]]:
30
+ ) -> dict[str, list[dr.genai.PromptTemplateVersion]]:
31
+ # Still missing in SDK
87
32
  prompt_template_versions_data = dr.utils.pagination.unpaginate(
88
33
  initial_url="genai/promptTemplates/versions/",
89
34
  initial_params={
@@ -94,35 +39,32 @@ def get_datarobot_prompt_template_versions(
94
39
  prompt_template_versions = defaultdict(list)
95
40
  for prompt_template_version in prompt_template_versions_data:
96
41
  prompt_template_versions[prompt_template_version["promptTemplateId"]].append(
97
- DrPromptVersion.from_dict(prompt_template_version)
42
+ dr.genai.PromptTemplateVersion(
43
+ id=prompt_template_version["id"],
44
+ prompt_template_id=prompt_template_version["promptTemplateId"],
45
+ prompt_text=prompt_template_version["promptText"],
46
+ commit_comment=prompt_template_version["commitComment"],
47
+ version=prompt_template_version["version"],
48
+ variables=prompt_template_version["variables"],
49
+ creation_date=prompt_template_version["creationDate"],
50
+ creation_user_id=prompt_template_version["creationUserId"],
51
+ user_name=prompt_template_version["userName"],
52
+ )
98
53
  )
99
54
  return prompt_template_versions
100
55
 
101
56
 
102
- def get_datarobot_prompt_template(prompt_template_id: str) -> DrPrompt | None:
103
- api_client = get_api_client()
57
+ def get_datarobot_prompt_template(prompt_template_id: str) -> dr.genai.PromptTemplate | None:
104
58
  try:
105
- prompt_template_response = api_client.get(
106
- f"genai/promptTemplates/{prompt_template_id}/", join_endpoint=True
107
- )
108
- prompt_template_json = prompt_template_response.json()
59
+ return dr.genai.PromptTemplate.get(prompt_template_id)
109
60
  except Exception:
110
61
  return None
111
62
 
112
- return DrPrompt.from_dict(prompt_template_json)
113
-
114
63
 
115
64
  def get_datarobot_prompt_template_version(
116
65
  prompt_template_id: str, prompt_template_version_id: str
117
- ) -> DrPromptVersion | None:
118
- api_client = get_api_client()
66
+ ) -> dr.genai.PromptTemplateVersion | None:
119
67
  try:
120
- prompt_template_version_response = api_client.get(
121
- f"genai/promptTemplates/{prompt_template_id}/versions/{prompt_template_version_id}/",
122
- join_endpoint=True,
123
- )
124
- prompt_template_version_json = prompt_template_version_response.json()
68
+ return dr.genai.PromptTemplateVersion.get(prompt_template_id, prompt_template_version_id)
125
69
  except Exception:
126
70
  return None
127
-
128
- return DrPromptVersion.from_dict(prompt_template_version_json)
@@ -18,15 +18,13 @@ from collections.abc import Callable
18
18
  from inspect import Parameter
19
19
  from inspect import Signature
20
20
 
21
+ import datarobot as dr
21
22
  from fastmcp.prompts.prompt import Prompt
22
23
  from pydantic import Field
23
24
 
24
25
  from datarobot_genai.drmcp.core.exceptions import DynamicPromptRegistrationError
25
26
  from datarobot_genai.drmcp.core.mcp_instance import register_prompt
26
27
 
27
- from .dr_lib import DrPrompt
28
- from .dr_lib import DrPromptVersion
29
- from .dr_lib import DrVariable
30
28
  from .dr_lib import get_datarobot_prompt_template_versions
31
29
  from .dr_lib import get_datarobot_prompt_templates
32
30
 
@@ -57,7 +55,8 @@ async def register_prompts_from_datarobot_prompt_management() -> None:
57
55
 
58
56
 
59
57
  async def register_prompt_from_datarobot_prompt_management(
60
- prompt_template: DrPrompt, prompt_template_version: DrPromptVersion | None = None
58
+ prompt_template: dr.genai.PromptTemplate,
59
+ prompt_template_version: dr.genai.PromptTemplateVersion | None = None,
61
60
  ) -> Prompt:
62
61
  """Register a single prompt.
63
62
 
@@ -173,7 +172,7 @@ def to_valid_mcp_prompt_name(s: str) -> str:
173
172
 
174
173
 
175
174
  def make_prompt_function(
176
- name: str, description: str, prompt_text: str, variables: list[DrVariable]
175
+ name: str, description: str, prompt_text: str, variables: list[dr.genai.Variable]
177
176
  ) -> Callable:
178
177
  params = []
179
178
  for v in variables:
@@ -22,6 +22,7 @@ from fastmcp import Context
22
22
  from fastmcp import FastMCP
23
23
  from fastmcp.exceptions import NotFoundError
24
24
  from fastmcp.prompts.prompt import Prompt
25
+ from fastmcp.server.dependencies import get_context
25
26
  from fastmcp.tools import FunctionTool
26
27
  from fastmcp.tools import Tool
27
28
  from fastmcp.utilities.types import NotSet
@@ -91,6 +92,34 @@ class TaggedFastMCP(FastMCP):
91
92
  self._deployments_map: dict[str, str] = {}
92
93
  self._prompts_map: dict[str, tuple[str, str]] = {}
93
94
 
95
+ async def notify_prompts_changed(self) -> None:
96
+ """
97
+ Notify connected clients that the prompt list has changed.
98
+
99
+ This method attempts to send a prompts/list_changed notification to inform
100
+ clients that they should refresh their prompt list.
101
+
102
+ Note: In stateless HTTP mode (default for this server), notifications may not
103
+ reach clients since each request is independent. This method still logs the
104
+ change for auditing purposes and will work if the server is configured for
105
+ stateful connections.
106
+
107
+ See: https://github.com/modelcontextprotocol/python-sdk/issues/710
108
+ """
109
+ logger.info("Prompt list changed - attempting to notify connected clients")
110
+
111
+ # Try to use FastMCP's built-in notification mechanism if in an MCP context
112
+ try:
113
+ context = get_context()
114
+ context._queue_prompt_list_changed()
115
+ logger.debug("Queued prompts_changed notification via MCP context")
116
+ except RuntimeError:
117
+ # No active MCP context - this is expected when called from REST API
118
+ logger.debug(
119
+ "No active MCP context for notification. "
120
+ "In stateless mode, clients will see changes on next request."
121
+ )
122
+
94
123
  @overload
95
124
  def tool(
96
125
  self,
@@ -286,6 +315,9 @@ class TaggedFastMCP(FastMCP):
286
315
  f"already mapped to {existing_prompt_template_version_id}. "
287
316
  f"Updating to version id = {prompt_template_version_id} and name = {prompt_name}"
288
317
  )
318
+ await self.remove_prompt_mapping(
319
+ prompt_template_id, existing_prompt_template_version_id
320
+ )
289
321
 
290
322
  self._prompts_map[prompt_template_id] = (prompt_template_version_id, prompt_name)
291
323
 
@@ -308,7 +340,7 @@ class TaggedFastMCP(FastMCP):
308
340
  f"skipping removal."
309
341
  )
310
342
  else:
311
- prompts_d = await mcp.get_prompts()
343
+ prompts_d = await self.get_prompts()
312
344
  for prompt in prompts_d.values():
313
345
  if (
314
346
  prompt.meta is not None
@@ -319,6 +351,9 @@ class TaggedFastMCP(FastMCP):
319
351
  prompt.disable()
320
352
 
321
353
  self._prompts_map.pop(prompt_template_id, None)
354
+
355
+ # Notify clients that the prompt list has changed
356
+ await self.notify_prompts_changed()
322
357
  else:
323
358
  logger.debug(
324
359
  f"Do not found prompt template with id = {prompt_template_id} in registry, "
@@ -526,17 +561,21 @@ async def register_prompt(
526
561
  )
527
562
 
528
563
  # Register the prompt
529
- registered_prompt = mcp.add_prompt(prompt)
530
564
  if prompt_template:
531
565
  prompt_template_id, prompt_template_version_id = prompt_template
532
566
  await mcp.set_prompt_mapping(
533
567
  prompt_template_id, prompt_template_version_id, prompt_name_no_duplicate
534
568
  )
535
569
 
570
+ registered_prompt = mcp.add_prompt(prompt)
571
+
536
572
  # Verify prompt is registered
537
573
  prompts = await mcp.get_prompts()
538
574
  if not any(prompt.name == prompt_name_no_duplicate for prompt in prompts.values()):
539
575
  raise RuntimeError(f"Prompt {prompt_name_no_duplicate} was not registered successfully")
540
576
  logger.info(f"Registered prompts: {len(prompts)}")
541
577
 
578
+ # Notify clients that the prompt list has changed
579
+ await mcp.notify_prompts_changed()
580
+
542
581
  return registered_prompt
@@ -428,7 +428,10 @@ def register_routes(mcp: TaggedFastMCP) -> None:
428
428
  """Refresh prompt templates."""
429
429
  try:
430
430
  await refresh_registered_prompt_template()
431
- return JSONResponse(status_code=HTTPStatus.NO_CONTENT, content=None)
431
+ return JSONResponse(
432
+ status_code=HTTPStatus.OK,
433
+ content={"message": "Prompts refreshed successfully"},
434
+ )
432
435
  except Exception as e:
433
436
  return JSONResponse(
434
437
  status_code=HTTPStatus.INTERNAL_SERVER_ERROR,
@@ -0,0 +1,95 @@
1
+ # Copyright 2025 DataRobot, Inc.
2
+ #
3
+ # Licensed under the Apache License, Version 2.0 (the "License");
4
+ # you may not use this file except in compliance with the License.
5
+ # You may obtain a copy of the License at
6
+ #
7
+ # http://www.apache.org/licenses/LICENSE-2.0
8
+ #
9
+ # Unless required by applicable law or agreed to in writing, software
10
+ # distributed under the License is distributed on an "AS IS" BASIS,
11
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ # See the License for the specific language governing permissions and
13
+ # limitations under the License.
14
+
15
+ """Tool configuration and enablement logic."""
16
+
17
+ from collections.abc import Callable
18
+ from enum import Enum
19
+ from typing import TYPE_CHECKING
20
+ from typing import TypedDict
21
+
22
+ if TYPE_CHECKING:
23
+ from .config import MCPServerConfig
24
+
25
+
26
+ class ToolType(str, Enum):
27
+ """Enumeration of available tool types."""
28
+
29
+ PREDICTIVE = "predictive"
30
+ JIRA = "jira"
31
+ CONFLUENCE = "confluence"
32
+
33
+
34
+ class ToolConfig(TypedDict):
35
+ """Configuration for a tool type."""
36
+
37
+ name: str
38
+ oauth_check: Callable[["MCPServerConfig"], bool] | None
39
+ directory: str
40
+ package_prefix: str
41
+ config_field_name: str # Name of the config field (e.g., "enable_predictive_tools")
42
+
43
+
44
+ # Tool configuration registry
45
+ TOOL_CONFIGS: dict[ToolType, ToolConfig] = {
46
+ ToolType.PREDICTIVE: ToolConfig(
47
+ name="predictive",
48
+ oauth_check=None,
49
+ directory="predictive",
50
+ package_prefix="datarobot_genai.drmcp.tools.predictive",
51
+ config_field_name="enable_predictive_tools",
52
+ ),
53
+ ToolType.JIRA: ToolConfig(
54
+ name="jira",
55
+ oauth_check=lambda config: config.is_jira_oauth_configured,
56
+ directory="jira",
57
+ package_prefix="datarobot_genai.drmcp.tools.jira",
58
+ config_field_name="enable_jira_tools",
59
+ ),
60
+ ToolType.CONFLUENCE: ToolConfig(
61
+ name="confluence",
62
+ oauth_check=lambda config: config.is_confluence_oauth_configured,
63
+ directory="confluence",
64
+ package_prefix="datarobot_genai.drmcp.tools.confluence",
65
+ config_field_name="enable_confluence_tools",
66
+ ),
67
+ }
68
+
69
+
70
+ def get_tool_enable_config_name(tool_type: ToolType) -> str:
71
+ """Get the configuration field name for enabling a tool."""
72
+ return TOOL_CONFIGS[tool_type]["config_field_name"]
73
+
74
+
75
+ def is_tool_enabled(tool_type: ToolType, config: "MCPServerConfig") -> bool:
76
+ """
77
+ Check if a tool is enabled based on configuration.
78
+
79
+ Args:
80
+ tool_type: The type of tool to check
81
+ config: The server configuration
82
+
83
+ Returns
84
+ -------
85
+ True if the tool is enabled, False otherwise
86
+ """
87
+ tool_config = TOOL_CONFIGS[tool_type]
88
+ enable_config_name = tool_config["config_field_name"]
89
+ is_enabled = getattr(config, enable_config_name)
90
+
91
+ # If tool is enabled, check OAuth requirements if needed
92
+ if is_enabled and tool_config["oauth_check"] is not None:
93
+ return tool_config["oauth_check"](config)
94
+
95
+ return is_enabled
@@ -16,6 +16,8 @@ import os
16
16
  from collections.abc import AsyncGenerator
17
17
  from contextlib import asynccontextmanager
18
18
 
19
+ import aiohttp
20
+ from aiohttp import ClientSession as HttpClientSession
19
21
  from mcp import ClientSession
20
22
  from mcp.client.streamable_http import streamablehttp_client
21
23
 
@@ -29,6 +31,11 @@ def get_dr_mcp_server_url() -> str | None:
29
31
  return os.environ.get("DR_MCP_SERVER_URL")
30
32
 
31
33
 
34
+ def get_dr_mcp_server_http_url() -> str | None:
35
+ """Get DataRobot MCP server http URL."""
36
+ return os.environ.get("DR_MCP_SERVER_HTTP_URL")
37
+
38
+
32
39
  def get_openai_llm_client_config() -> dict[str, str]:
33
40
  """Get OpenAI LLM client configuration."""
34
41
  openai_api_key = os.environ.get("OPENAI_API_KEY")
@@ -94,3 +101,25 @@ async def ete_test_mcp_session(
94
101
  yield session
95
102
  except asyncio.TimeoutError:
96
103
  raise TimeoutError(f"Check if the MCP server is running at {get_dr_mcp_server_url()}")
104
+
105
+
106
+ @asynccontextmanager
107
+ async def ete_test_http_session(
108
+ additional_headers: dict[str, str] | None = None,
109
+ ) -> AsyncGenerator[HttpClientSession, None]:
110
+ """Create an HTTP session for each test that can connect to MCP custom http routes.
111
+
112
+ Parameters
113
+ ----------
114
+ additional_headers : dict[str, str], optional
115
+ Additional headers to include in the HTTP session (e.g., auth headers for testing).
116
+ """
117
+ headers = get_headers()
118
+ if additional_headers:
119
+ headers.update(additional_headers)
120
+
121
+ async with ete_test_mcp_session(additional_headers=additional_headers):
122
+ async with aiohttp.ClientSession(
123
+ base_url=get_dr_mcp_server_http_url(), headers=headers
124
+ ) as client:
125
+ yield client