datarobot-genai 0.1.75__py3-none-any.whl → 0.2.11__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- datarobot_genai/core/agents/base.py +2 -1
- datarobot_genai/core/chat/responses.py +32 -4
- datarobot_genai/drmcp/core/config.py +52 -0
- datarobot_genai/drmcp/core/dr_mcp_server.py +45 -8
- datarobot_genai/drmcp/core/dynamic_prompts/dr_lib.py +22 -80
- datarobot_genai/drmcp/core/dynamic_prompts/register.py +4 -5
- datarobot_genai/drmcp/core/mcp_instance.py +41 -2
- datarobot_genai/drmcp/core/routes.py +4 -1
- datarobot_genai/drmcp/core/tool_config.py +95 -0
- datarobot_genai/drmcp/test_utils/mcp_utils_ete.py +29 -0
- datarobot_genai/drmcp/test_utils/openai_llm_mcp_client.py +6 -1
- datarobot_genai/drmcp/tools/clients/__init__.py +14 -0
- datarobot_genai/drmcp/tools/clients/atlassian.py +188 -0
- datarobot_genai/drmcp/tools/clients/confluence.py +196 -0
- datarobot_genai/drmcp/tools/clients/jira.py +147 -0
- datarobot_genai/drmcp/tools/clients/s3.py +28 -0
- datarobot_genai/drmcp/tools/confluence/__init__.py +14 -0
- datarobot_genai/drmcp/tools/confluence/tools.py +81 -0
- datarobot_genai/drmcp/tools/jira/__init__.py +14 -0
- datarobot_genai/drmcp/tools/jira/tools.py +52 -0
- datarobot_genai/drmcp/tools/predictive/predict.py +1 -1
- datarobot_genai/drmcp/tools/predictive/predict_realtime.py +1 -1
- datarobot_genai/langgraph/agent.py +143 -42
- datarobot_genai/nat/agent.py +4 -0
- datarobot_genai/nat/datarobot_auth_provider.py +110 -0
- datarobot_genai/nat/datarobot_mcp_client.py +234 -0
- {datarobot_genai-0.1.75.dist-info → datarobot_genai-0.2.11.dist-info}/METADATA +9 -2
- {datarobot_genai-0.1.75.dist-info → datarobot_genai-0.2.11.dist-info}/RECORD +32 -20
- {datarobot_genai-0.1.75.dist-info → datarobot_genai-0.2.11.dist-info}/entry_points.txt +2 -0
- {datarobot_genai-0.1.75.dist-info → datarobot_genai-0.2.11.dist-info}/WHEEL +0 -0
- {datarobot_genai-0.1.75.dist-info → datarobot_genai-0.2.11.dist-info}/licenses/AUTHORS +0 -0
- {datarobot_genai-0.1.75.dist-info → datarobot_genai-0.2.11.dist-info}/licenses/LICENSE +0 -0
|
@@ -23,6 +23,7 @@ from typing import TypedDict
|
|
|
23
23
|
from typing import TypeVar
|
|
24
24
|
from typing import cast
|
|
25
25
|
|
|
26
|
+
from ag_ui.core import Event
|
|
26
27
|
from openai.types.chat import CompletionCreateParams
|
|
27
28
|
from ragas import MultiTurnSample
|
|
28
29
|
|
|
@@ -167,7 +168,7 @@ class UsageMetrics(TypedDict):
|
|
|
167
168
|
|
|
168
169
|
# Canonical return type for DRUM-compatible invoke implementations
|
|
169
170
|
InvokeReturn = (
|
|
170
|
-
AsyncGenerator[tuple[str, MultiTurnSample | None, UsageMetrics], None]
|
|
171
|
+
AsyncGenerator[tuple[str | Event, MultiTurnSample | None, UsageMetrics], None]
|
|
171
172
|
| tuple[str, MultiTurnSample | None, UsageMetrics]
|
|
172
173
|
)
|
|
173
174
|
|
|
@@ -27,6 +27,10 @@ from concurrent.futures import ThreadPoolExecutor
|
|
|
27
27
|
from typing import Any
|
|
28
28
|
from typing import TypeVar
|
|
29
29
|
|
|
30
|
+
from ag_ui.core import BaseEvent
|
|
31
|
+
from ag_ui.core import Event
|
|
32
|
+
from ag_ui.core import TextMessageChunkEvent
|
|
33
|
+
from ag_ui.core import TextMessageContentEvent
|
|
30
34
|
from openai.types import CompletionUsage
|
|
31
35
|
from openai.types.chat import ChatCompletion
|
|
32
36
|
from openai.types.chat import ChatCompletionChunk
|
|
@@ -45,6 +49,7 @@ class CustomModelChatResponse(ChatCompletion):
|
|
|
45
49
|
|
|
46
50
|
class CustomModelStreamingResponse(ChatCompletionChunk):
|
|
47
51
|
pipeline_interactions: str | None = None
|
|
52
|
+
event: Event | None = None
|
|
48
53
|
|
|
49
54
|
|
|
50
55
|
def to_custom_model_chat_response(
|
|
@@ -88,7 +93,7 @@ def to_custom_model_streaming_response(
|
|
|
88
93
|
thread_pool_executor: ThreadPoolExecutor,
|
|
89
94
|
event_loop: AbstractEventLoop,
|
|
90
95
|
streaming_response_generator: AsyncGenerator[
|
|
91
|
-
tuple[str, MultiTurnSample | None, dict[str, int]], None
|
|
96
|
+
tuple[str | Event, MultiTurnSample | None, dict[str, int]], None
|
|
92
97
|
],
|
|
93
98
|
model: str | object | None,
|
|
94
99
|
) -> Iterator[CustomModelStreamingResponse]:
|
|
@@ -110,7 +115,7 @@ def to_custom_model_streaming_response(
|
|
|
110
115
|
while True:
|
|
111
116
|
try:
|
|
112
117
|
(
|
|
113
|
-
|
|
118
|
+
response_text_or_event,
|
|
114
119
|
pipeline_interactions,
|
|
115
120
|
usage_metrics,
|
|
116
121
|
) = thread_pool_executor.submit(
|
|
@@ -119,10 +124,10 @@ def to_custom_model_streaming_response(
|
|
|
119
124
|
last_pipeline_interactions = pipeline_interactions
|
|
120
125
|
last_usage_metrics = usage_metrics
|
|
121
126
|
|
|
122
|
-
if
|
|
127
|
+
if isinstance(response_text_or_event, str) and response_text_or_event:
|
|
123
128
|
choice = ChunkChoice(
|
|
124
129
|
index=0,
|
|
125
|
-
delta=ChoiceDelta(role="assistant", content=
|
|
130
|
+
delta=ChoiceDelta(role="assistant", content=response_text_or_event),
|
|
126
131
|
finish_reason=None,
|
|
127
132
|
)
|
|
128
133
|
yield CustomModelStreamingResponse(
|
|
@@ -135,6 +140,29 @@ def to_custom_model_streaming_response(
|
|
|
135
140
|
if usage_metrics
|
|
136
141
|
else None,
|
|
137
142
|
)
|
|
143
|
+
elif isinstance(response_text_or_event, BaseEvent):
|
|
144
|
+
content = ""
|
|
145
|
+
if isinstance(
|
|
146
|
+
response_text_or_event, (TextMessageContentEvent, TextMessageChunkEvent)
|
|
147
|
+
):
|
|
148
|
+
content = response_text_or_event.delta or content
|
|
149
|
+
choice = ChunkChoice(
|
|
150
|
+
index=0,
|
|
151
|
+
delta=ChoiceDelta(role="assistant", content=content),
|
|
152
|
+
finish_reason=None,
|
|
153
|
+
)
|
|
154
|
+
|
|
155
|
+
yield CustomModelStreamingResponse(
|
|
156
|
+
id=completion_id,
|
|
157
|
+
object="chat.completion.chunk",
|
|
158
|
+
created=created,
|
|
159
|
+
model=model,
|
|
160
|
+
choices=[choice],
|
|
161
|
+
usage=CompletionUsage.model_validate(required_usage_metrics | usage_metrics)
|
|
162
|
+
if usage_metrics
|
|
163
|
+
else None,
|
|
164
|
+
event=response_text_or_event,
|
|
165
|
+
)
|
|
138
166
|
except StopAsyncIteration:
|
|
139
167
|
break
|
|
140
168
|
event_loop.run_until_complete(streaming_response_generator.aclose())
|
|
@@ -197,6 +197,54 @@ class MCPServerConfig(BaseSettings):
|
|
|
197
197
|
description="Enable/disable predictive tools",
|
|
198
198
|
)
|
|
199
199
|
|
|
200
|
+
# Jira tools
|
|
201
|
+
enable_jira_tools: bool = Field(
|
|
202
|
+
default=False,
|
|
203
|
+
validation_alias=AliasChoices(
|
|
204
|
+
RUNTIME_PARAM_ENV_VAR_NAME_PREFIX + "ENABLE_JIRA_TOOLS",
|
|
205
|
+
"ENABLE_JIRA_TOOLS",
|
|
206
|
+
),
|
|
207
|
+
description="Enable/disable Jira tools",
|
|
208
|
+
)
|
|
209
|
+
is_jira_oauth_provider_configured: bool = Field(
|
|
210
|
+
default=False,
|
|
211
|
+
validation_alias=AliasChoices(
|
|
212
|
+
RUNTIME_PARAM_ENV_VAR_NAME_PREFIX + "IS_JIRA_OAUTH_PROVIDER_CONFIGURED",
|
|
213
|
+
"IS_JIRA_OAUTH_PROVIDER_CONFIGURED",
|
|
214
|
+
),
|
|
215
|
+
description="Whether Jira OAuth provider is configured for Jira integration",
|
|
216
|
+
)
|
|
217
|
+
|
|
218
|
+
@property
|
|
219
|
+
def is_jira_oauth_configured(self) -> bool:
|
|
220
|
+
return self.is_jira_oauth_provider_configured or bool(
|
|
221
|
+
os.getenv("JIRA_CLIENT_ID") and os.getenv("JIRA_CLIENT_SECRET")
|
|
222
|
+
)
|
|
223
|
+
|
|
224
|
+
# Confluence tools
|
|
225
|
+
enable_confluence_tools: bool = Field(
|
|
226
|
+
default=False,
|
|
227
|
+
validation_alias=AliasChoices(
|
|
228
|
+
RUNTIME_PARAM_ENV_VAR_NAME_PREFIX + "ENABLE_CONFLUENCE_TOOLS",
|
|
229
|
+
"ENABLE_CONFLUENCE_TOOLS",
|
|
230
|
+
),
|
|
231
|
+
description="Enable/disable Confluence tools",
|
|
232
|
+
)
|
|
233
|
+
is_confluence_oauth_provider_configured: bool = Field(
|
|
234
|
+
default=False,
|
|
235
|
+
validation_alias=AliasChoices(
|
|
236
|
+
RUNTIME_PARAM_ENV_VAR_NAME_PREFIX + "IS_CONFLUENCE_OAUTH_PROVIDER_CONFIGURED",
|
|
237
|
+
"IS_CONFLUENCE_OAUTH_PROVIDER_CONFIGURED",
|
|
238
|
+
),
|
|
239
|
+
description="Whether Confluence OAuth provider is configured for Confluence integration",
|
|
240
|
+
)
|
|
241
|
+
|
|
242
|
+
@property
|
|
243
|
+
def is_confluence_oauth_configured(self) -> bool:
|
|
244
|
+
return self.is_confluence_oauth_provider_configured or bool(
|
|
245
|
+
os.getenv("CONFLUENCE_CLIENT_ID") and os.getenv("CONFLUENCE_CLIENT_SECRET")
|
|
246
|
+
)
|
|
247
|
+
|
|
200
248
|
@field_validator(
|
|
201
249
|
"otel_attributes",
|
|
202
250
|
mode="before",
|
|
@@ -220,6 +268,10 @@ class MCPServerConfig(BaseSettings):
|
|
|
220
268
|
"tool_registration_duplicate_behavior",
|
|
221
269
|
"mcp_server_register_dynamic_prompts_on_startup",
|
|
222
270
|
"enable_predictive_tools",
|
|
271
|
+
"enable_jira_tools",
|
|
272
|
+
"is_jira_oauth_provider_configured",
|
|
273
|
+
"enable_confluence_tools",
|
|
274
|
+
"is_confluence_oauth_provider_configured",
|
|
223
275
|
mode="before",
|
|
224
276
|
)
|
|
225
277
|
@classmethod
|
|
@@ -40,6 +40,8 @@ from .routes_utils import prefix_mount_path
|
|
|
40
40
|
from .server_life_cycle import BaseServerLifecycle
|
|
41
41
|
from .telemetry import OtelASGIMiddleware
|
|
42
42
|
from .telemetry import initialize_telemetry
|
|
43
|
+
from .tool_config import TOOL_CONFIGS
|
|
44
|
+
from .tool_config import is_tool_enabled
|
|
43
45
|
|
|
44
46
|
|
|
45
47
|
def _import_modules_from_dir(
|
|
@@ -115,6 +117,9 @@ class DataRobotMCPServer:
|
|
|
115
117
|
self._mcp = mcp
|
|
116
118
|
self._mcp_transport = transport
|
|
117
119
|
|
|
120
|
+
# Configure MCP server capabilities
|
|
121
|
+
self._configure_mcp_capabilities()
|
|
122
|
+
|
|
118
123
|
# Initialize telemetry
|
|
119
124
|
initialize_telemetry(mcp)
|
|
120
125
|
|
|
@@ -139,11 +144,12 @@ class DataRobotMCPServer:
|
|
|
139
144
|
|
|
140
145
|
# Load static tools modules
|
|
141
146
|
base_dir = os.path.dirname(os.path.dirname(__file__))
|
|
142
|
-
|
|
143
|
-
|
|
144
|
-
|
|
145
|
-
|
|
146
|
-
|
|
147
|
+
for tool_type, tool_config in TOOL_CONFIGS.items():
|
|
148
|
+
if is_tool_enabled(tool_type, self._config):
|
|
149
|
+
_import_modules_from_dir(
|
|
150
|
+
os.path.join(base_dir, "tools", tool_config["directory"]),
|
|
151
|
+
tool_config["package_prefix"],
|
|
152
|
+
)
|
|
147
153
|
|
|
148
154
|
# Load memory management tools if available
|
|
149
155
|
if self._memory_manager:
|
|
@@ -163,6 +169,37 @@ class DataRobotMCPServer:
|
|
|
163
169
|
if transport == "streamable-http":
|
|
164
170
|
register_routes(self._mcp)
|
|
165
171
|
|
|
172
|
+
def _configure_mcp_capabilities(self) -> None:
|
|
173
|
+
"""Configure MCP capabilities that FastMCP doesn't expose directly.
|
|
174
|
+
|
|
175
|
+
See: https://github.com/modelcontextprotocol/python-sdk/issues/1126
|
|
176
|
+
"""
|
|
177
|
+
server = self._mcp._mcp_server
|
|
178
|
+
|
|
179
|
+
# Declare prompts_changed capability (capabilities.prompts.listChanged: true)
|
|
180
|
+
server.notification_options.prompts_changed = True
|
|
181
|
+
|
|
182
|
+
# Declare experimental capabilities ( experimental.dynamic_prompts: true)
|
|
183
|
+
server.experimental_capabilities = {"dynamic_prompts": {"enabled": True}}
|
|
184
|
+
|
|
185
|
+
# Patch to include experimental_capabilities (FastMCP doesn't expose this)
|
|
186
|
+
original = server.create_initialization_options
|
|
187
|
+
|
|
188
|
+
def patched(
|
|
189
|
+
notification_options: Any = None,
|
|
190
|
+
experimental_capabilities: dict[str, dict[str, Any]] | None = None,
|
|
191
|
+
**kwargs: Any,
|
|
192
|
+
) -> Any:
|
|
193
|
+
if experimental_capabilities is None:
|
|
194
|
+
experimental_capabilities = getattr(server, "experimental_capabilities", None)
|
|
195
|
+
return original(
|
|
196
|
+
notification_options=notification_options,
|
|
197
|
+
experimental_capabilities=experimental_capabilities,
|
|
198
|
+
**kwargs,
|
|
199
|
+
)
|
|
200
|
+
|
|
201
|
+
server.create_initialization_options = patched
|
|
202
|
+
|
|
166
203
|
def run(self, show_banner: bool = False) -> None:
|
|
167
204
|
"""Run the DataRobot MCP server synchronously."""
|
|
168
205
|
try:
|
|
@@ -179,6 +216,9 @@ class DataRobotMCPServer:
|
|
|
179
216
|
self._logger.info("Registering dynamic prompts from prompt management...")
|
|
180
217
|
asyncio.run(register_prompts_from_datarobot_prompt_management())
|
|
181
218
|
|
|
219
|
+
# Execute pre-server start actions
|
|
220
|
+
asyncio.run(self._lifecycle.pre_server_start(self._mcp))
|
|
221
|
+
|
|
182
222
|
# List registered tools, prompts, and resources before starting server
|
|
183
223
|
tools = asyncio.run(self._mcp._list_tools_mcp())
|
|
184
224
|
prompts = asyncio.run(self._mcp._list_prompts_mcp())
|
|
@@ -198,9 +238,6 @@ class DataRobotMCPServer:
|
|
|
198
238
|
for resource in resources:
|
|
199
239
|
self._logger.info(f" > {resource.name}")
|
|
200
240
|
|
|
201
|
-
# Execute pre-server start actions
|
|
202
|
-
asyncio.run(self._lifecycle.pre_server_start(self._mcp))
|
|
203
|
-
|
|
204
241
|
# Create event loop for async operations
|
|
205
242
|
loop = asyncio.new_event_loop()
|
|
206
243
|
asyncio.set_event_loop(loop)
|
|
@@ -12,78 +12,23 @@
|
|
|
12
12
|
# See the License for the specific language governing permissions and
|
|
13
13
|
# limitations under the License.
|
|
14
14
|
from collections import defaultdict
|
|
15
|
-
from dataclasses import dataclass
|
|
16
15
|
|
|
17
16
|
import datarobot as dr
|
|
18
17
|
|
|
19
18
|
from datarobot_genai.drmcp.core.clients import get_api_client
|
|
20
19
|
|
|
21
|
-
# Needed SDK version (3.10.0) is not published yet. We'll reimplement simplified version of it.
|
|
22
|
-
# get_datarobot_prompt_templates = dr.genai.PromptTemplate.list()
|
|
23
|
-
# DrPrompt = dr.genai.PromptTemplate
|
|
24
|
-
# DrPromptVersion = dr.genai.PromptTemplateVersion
|
|
25
|
-
# DrVariable = dr.genai.Variable
|
|
26
20
|
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
@dataclass
|
|
35
|
-
class DrPromptVersion:
|
|
36
|
-
id: str
|
|
37
|
-
prompt_template_id: str
|
|
38
|
-
version: int
|
|
39
|
-
prompt_text: str
|
|
40
|
-
variables: list[DrVariable]
|
|
41
|
-
|
|
42
|
-
@classmethod
|
|
43
|
-
def from_dict(cls, d: dict) -> "DrPromptVersion":
|
|
44
|
-
variables = [
|
|
45
|
-
DrVariable(name=v["name"], description=v["description"]) for v in d["variables"]
|
|
46
|
-
]
|
|
47
|
-
return cls(
|
|
48
|
-
id=d["id"],
|
|
49
|
-
prompt_template_id=d["promptTemplateId"],
|
|
50
|
-
version=d["version"],
|
|
51
|
-
prompt_text=d["promptText"],
|
|
52
|
-
variables=variables,
|
|
53
|
-
)
|
|
54
|
-
|
|
55
|
-
|
|
56
|
-
@dataclass
|
|
57
|
-
class DrPrompt:
|
|
58
|
-
id: str
|
|
59
|
-
name: str
|
|
60
|
-
description: str
|
|
61
|
-
|
|
62
|
-
def get_latest_version(self) -> DrPromptVersion | None:
|
|
63
|
-
all_prompt_template_versions = get_datarobot_prompt_template_versions([self.id])
|
|
64
|
-
prompt_template_versions = all_prompt_template_versions.get(self.id)
|
|
65
|
-
|
|
66
|
-
if not prompt_template_versions:
|
|
67
|
-
return None
|
|
68
|
-
latest_version = max(prompt_template_versions, key=lambda v: v.version)
|
|
69
|
-
return latest_version
|
|
70
|
-
|
|
71
|
-
@classmethod
|
|
72
|
-
def from_dict(cls, d: dict) -> "DrPrompt":
|
|
73
|
-
return cls(id=d["id"], name=d["name"], description=d["description"])
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
def get_datarobot_prompt_templates() -> list[DrPrompt]:
|
|
77
|
-
prompt_templates_data = dr.utils.pagination.unpaginate(
|
|
78
|
-
initial_url="genai/promptTemplates/", initial_params={}, client=get_api_client()
|
|
79
|
-
)
|
|
80
|
-
|
|
81
|
-
return [DrPrompt.from_dict(prompt_template) for prompt_template in prompt_templates_data]
|
|
21
|
+
def get_datarobot_prompt_templates() -> list[dr.genai.PromptTemplate]:
|
|
22
|
+
try:
|
|
23
|
+
return dr.genai.PromptTemplate.list()
|
|
24
|
+
except Exception:
|
|
25
|
+
return []
|
|
82
26
|
|
|
83
27
|
|
|
84
28
|
def get_datarobot_prompt_template_versions(
|
|
85
29
|
prompt_template_ids: list[str],
|
|
86
|
-
) -> dict[str, list[
|
|
30
|
+
) -> dict[str, list[dr.genai.PromptTemplateVersion]]:
|
|
31
|
+
# Still missing in SDK
|
|
87
32
|
prompt_template_versions_data = dr.utils.pagination.unpaginate(
|
|
88
33
|
initial_url="genai/promptTemplates/versions/",
|
|
89
34
|
initial_params={
|
|
@@ -94,35 +39,32 @@ def get_datarobot_prompt_template_versions(
|
|
|
94
39
|
prompt_template_versions = defaultdict(list)
|
|
95
40
|
for prompt_template_version in prompt_template_versions_data:
|
|
96
41
|
prompt_template_versions[prompt_template_version["promptTemplateId"]].append(
|
|
97
|
-
|
|
42
|
+
dr.genai.PromptTemplateVersion(
|
|
43
|
+
id=prompt_template_version["id"],
|
|
44
|
+
prompt_template_id=prompt_template_version["promptTemplateId"],
|
|
45
|
+
prompt_text=prompt_template_version["promptText"],
|
|
46
|
+
commit_comment=prompt_template_version["commitComment"],
|
|
47
|
+
version=prompt_template_version["version"],
|
|
48
|
+
variables=prompt_template_version["variables"],
|
|
49
|
+
creation_date=prompt_template_version["creationDate"],
|
|
50
|
+
creation_user_id=prompt_template_version["creationUserId"],
|
|
51
|
+
user_name=prompt_template_version["userName"],
|
|
52
|
+
)
|
|
98
53
|
)
|
|
99
54
|
return prompt_template_versions
|
|
100
55
|
|
|
101
56
|
|
|
102
|
-
def get_datarobot_prompt_template(prompt_template_id: str) ->
|
|
103
|
-
api_client = get_api_client()
|
|
57
|
+
def get_datarobot_prompt_template(prompt_template_id: str) -> dr.genai.PromptTemplate | None:
|
|
104
58
|
try:
|
|
105
|
-
|
|
106
|
-
f"genai/promptTemplates/{prompt_template_id}/", join_endpoint=True
|
|
107
|
-
)
|
|
108
|
-
prompt_template_json = prompt_template_response.json()
|
|
59
|
+
return dr.genai.PromptTemplate.get(prompt_template_id)
|
|
109
60
|
except Exception:
|
|
110
61
|
return None
|
|
111
62
|
|
|
112
|
-
return DrPrompt.from_dict(prompt_template_json)
|
|
113
|
-
|
|
114
63
|
|
|
115
64
|
def get_datarobot_prompt_template_version(
|
|
116
65
|
prompt_template_id: str, prompt_template_version_id: str
|
|
117
|
-
) ->
|
|
118
|
-
api_client = get_api_client()
|
|
66
|
+
) -> dr.genai.PromptTemplateVersion | None:
|
|
119
67
|
try:
|
|
120
|
-
|
|
121
|
-
f"genai/promptTemplates/{prompt_template_id}/versions/{prompt_template_version_id}/",
|
|
122
|
-
join_endpoint=True,
|
|
123
|
-
)
|
|
124
|
-
prompt_template_version_json = prompt_template_version_response.json()
|
|
68
|
+
return dr.genai.PromptTemplateVersion.get(prompt_template_id, prompt_template_version_id)
|
|
125
69
|
except Exception:
|
|
126
70
|
return None
|
|
127
|
-
|
|
128
|
-
return DrPromptVersion.from_dict(prompt_template_version_json)
|
|
@@ -18,15 +18,13 @@ from collections.abc import Callable
|
|
|
18
18
|
from inspect import Parameter
|
|
19
19
|
from inspect import Signature
|
|
20
20
|
|
|
21
|
+
import datarobot as dr
|
|
21
22
|
from fastmcp.prompts.prompt import Prompt
|
|
22
23
|
from pydantic import Field
|
|
23
24
|
|
|
24
25
|
from datarobot_genai.drmcp.core.exceptions import DynamicPromptRegistrationError
|
|
25
26
|
from datarobot_genai.drmcp.core.mcp_instance import register_prompt
|
|
26
27
|
|
|
27
|
-
from .dr_lib import DrPrompt
|
|
28
|
-
from .dr_lib import DrPromptVersion
|
|
29
|
-
from .dr_lib import DrVariable
|
|
30
28
|
from .dr_lib import get_datarobot_prompt_template_versions
|
|
31
29
|
from .dr_lib import get_datarobot_prompt_templates
|
|
32
30
|
|
|
@@ -57,7 +55,8 @@ async def register_prompts_from_datarobot_prompt_management() -> None:
|
|
|
57
55
|
|
|
58
56
|
|
|
59
57
|
async def register_prompt_from_datarobot_prompt_management(
|
|
60
|
-
prompt_template:
|
|
58
|
+
prompt_template: dr.genai.PromptTemplate,
|
|
59
|
+
prompt_template_version: dr.genai.PromptTemplateVersion | None = None,
|
|
61
60
|
) -> Prompt:
|
|
62
61
|
"""Register a single prompt.
|
|
63
62
|
|
|
@@ -173,7 +172,7 @@ def to_valid_mcp_prompt_name(s: str) -> str:
|
|
|
173
172
|
|
|
174
173
|
|
|
175
174
|
def make_prompt_function(
|
|
176
|
-
name: str, description: str, prompt_text: str, variables: list[
|
|
175
|
+
name: str, description: str, prompt_text: str, variables: list[dr.genai.Variable]
|
|
177
176
|
) -> Callable:
|
|
178
177
|
params = []
|
|
179
178
|
for v in variables:
|
|
@@ -22,6 +22,7 @@ from fastmcp import Context
|
|
|
22
22
|
from fastmcp import FastMCP
|
|
23
23
|
from fastmcp.exceptions import NotFoundError
|
|
24
24
|
from fastmcp.prompts.prompt import Prompt
|
|
25
|
+
from fastmcp.server.dependencies import get_context
|
|
25
26
|
from fastmcp.tools import FunctionTool
|
|
26
27
|
from fastmcp.tools import Tool
|
|
27
28
|
from fastmcp.utilities.types import NotSet
|
|
@@ -91,6 +92,34 @@ class TaggedFastMCP(FastMCP):
|
|
|
91
92
|
self._deployments_map: dict[str, str] = {}
|
|
92
93
|
self._prompts_map: dict[str, tuple[str, str]] = {}
|
|
93
94
|
|
|
95
|
+
async def notify_prompts_changed(self) -> None:
|
|
96
|
+
"""
|
|
97
|
+
Notify connected clients that the prompt list has changed.
|
|
98
|
+
|
|
99
|
+
This method attempts to send a prompts/list_changed notification to inform
|
|
100
|
+
clients that they should refresh their prompt list.
|
|
101
|
+
|
|
102
|
+
Note: In stateless HTTP mode (default for this server), notifications may not
|
|
103
|
+
reach clients since each request is independent. This method still logs the
|
|
104
|
+
change for auditing purposes and will work if the server is configured for
|
|
105
|
+
stateful connections.
|
|
106
|
+
|
|
107
|
+
See: https://github.com/modelcontextprotocol/python-sdk/issues/710
|
|
108
|
+
"""
|
|
109
|
+
logger.info("Prompt list changed - attempting to notify connected clients")
|
|
110
|
+
|
|
111
|
+
# Try to use FastMCP's built-in notification mechanism if in an MCP context
|
|
112
|
+
try:
|
|
113
|
+
context = get_context()
|
|
114
|
+
context._queue_prompt_list_changed()
|
|
115
|
+
logger.debug("Queued prompts_changed notification via MCP context")
|
|
116
|
+
except RuntimeError:
|
|
117
|
+
# No active MCP context - this is expected when called from REST API
|
|
118
|
+
logger.debug(
|
|
119
|
+
"No active MCP context for notification. "
|
|
120
|
+
"In stateless mode, clients will see changes on next request."
|
|
121
|
+
)
|
|
122
|
+
|
|
94
123
|
@overload
|
|
95
124
|
def tool(
|
|
96
125
|
self,
|
|
@@ -286,6 +315,9 @@ class TaggedFastMCP(FastMCP):
|
|
|
286
315
|
f"already mapped to {existing_prompt_template_version_id}. "
|
|
287
316
|
f"Updating to version id = {prompt_template_version_id} and name = {prompt_name}"
|
|
288
317
|
)
|
|
318
|
+
await self.remove_prompt_mapping(
|
|
319
|
+
prompt_template_id, existing_prompt_template_version_id
|
|
320
|
+
)
|
|
289
321
|
|
|
290
322
|
self._prompts_map[prompt_template_id] = (prompt_template_version_id, prompt_name)
|
|
291
323
|
|
|
@@ -308,7 +340,7 @@ class TaggedFastMCP(FastMCP):
|
|
|
308
340
|
f"skipping removal."
|
|
309
341
|
)
|
|
310
342
|
else:
|
|
311
|
-
prompts_d = await
|
|
343
|
+
prompts_d = await self.get_prompts()
|
|
312
344
|
for prompt in prompts_d.values():
|
|
313
345
|
if (
|
|
314
346
|
prompt.meta is not None
|
|
@@ -319,6 +351,9 @@ class TaggedFastMCP(FastMCP):
|
|
|
319
351
|
prompt.disable()
|
|
320
352
|
|
|
321
353
|
self._prompts_map.pop(prompt_template_id, None)
|
|
354
|
+
|
|
355
|
+
# Notify clients that the prompt list has changed
|
|
356
|
+
await self.notify_prompts_changed()
|
|
322
357
|
else:
|
|
323
358
|
logger.debug(
|
|
324
359
|
f"Do not found prompt template with id = {prompt_template_id} in registry, "
|
|
@@ -526,17 +561,21 @@ async def register_prompt(
|
|
|
526
561
|
)
|
|
527
562
|
|
|
528
563
|
# Register the prompt
|
|
529
|
-
registered_prompt = mcp.add_prompt(prompt)
|
|
530
564
|
if prompt_template:
|
|
531
565
|
prompt_template_id, prompt_template_version_id = prompt_template
|
|
532
566
|
await mcp.set_prompt_mapping(
|
|
533
567
|
prompt_template_id, prompt_template_version_id, prompt_name_no_duplicate
|
|
534
568
|
)
|
|
535
569
|
|
|
570
|
+
registered_prompt = mcp.add_prompt(prompt)
|
|
571
|
+
|
|
536
572
|
# Verify prompt is registered
|
|
537
573
|
prompts = await mcp.get_prompts()
|
|
538
574
|
if not any(prompt.name == prompt_name_no_duplicate for prompt in prompts.values()):
|
|
539
575
|
raise RuntimeError(f"Prompt {prompt_name_no_duplicate} was not registered successfully")
|
|
540
576
|
logger.info(f"Registered prompts: {len(prompts)}")
|
|
541
577
|
|
|
578
|
+
# Notify clients that the prompt list has changed
|
|
579
|
+
await mcp.notify_prompts_changed()
|
|
580
|
+
|
|
542
581
|
return registered_prompt
|
|
@@ -428,7 +428,10 @@ def register_routes(mcp: TaggedFastMCP) -> None:
|
|
|
428
428
|
"""Refresh prompt templates."""
|
|
429
429
|
try:
|
|
430
430
|
await refresh_registered_prompt_template()
|
|
431
|
-
return JSONResponse(
|
|
431
|
+
return JSONResponse(
|
|
432
|
+
status_code=HTTPStatus.OK,
|
|
433
|
+
content={"message": "Prompts refreshed successfully"},
|
|
434
|
+
)
|
|
432
435
|
except Exception as e:
|
|
433
436
|
return JSONResponse(
|
|
434
437
|
status_code=HTTPStatus.INTERNAL_SERVER_ERROR,
|
|
@@ -0,0 +1,95 @@
|
|
|
1
|
+
# Copyright 2025 DataRobot, Inc.
|
|
2
|
+
#
|
|
3
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
|
+
# you may not use this file except in compliance with the License.
|
|
5
|
+
# You may obtain a copy of the License at
|
|
6
|
+
#
|
|
7
|
+
# http://www.apache.org/licenses/LICENSE-2.0
|
|
8
|
+
#
|
|
9
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
10
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
11
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
12
|
+
# See the License for the specific language governing permissions and
|
|
13
|
+
# limitations under the License.
|
|
14
|
+
|
|
15
|
+
"""Tool configuration and enablement logic."""
|
|
16
|
+
|
|
17
|
+
from collections.abc import Callable
|
|
18
|
+
from enum import Enum
|
|
19
|
+
from typing import TYPE_CHECKING
|
|
20
|
+
from typing import TypedDict
|
|
21
|
+
|
|
22
|
+
if TYPE_CHECKING:
|
|
23
|
+
from .config import MCPServerConfig
|
|
24
|
+
|
|
25
|
+
|
|
26
|
+
class ToolType(str, Enum):
|
|
27
|
+
"""Enumeration of available tool types."""
|
|
28
|
+
|
|
29
|
+
PREDICTIVE = "predictive"
|
|
30
|
+
JIRA = "jira"
|
|
31
|
+
CONFLUENCE = "confluence"
|
|
32
|
+
|
|
33
|
+
|
|
34
|
+
class ToolConfig(TypedDict):
|
|
35
|
+
"""Configuration for a tool type."""
|
|
36
|
+
|
|
37
|
+
name: str
|
|
38
|
+
oauth_check: Callable[["MCPServerConfig"], bool] | None
|
|
39
|
+
directory: str
|
|
40
|
+
package_prefix: str
|
|
41
|
+
config_field_name: str # Name of the config field (e.g., "enable_predictive_tools")
|
|
42
|
+
|
|
43
|
+
|
|
44
|
+
# Tool configuration registry
|
|
45
|
+
TOOL_CONFIGS: dict[ToolType, ToolConfig] = {
|
|
46
|
+
ToolType.PREDICTIVE: ToolConfig(
|
|
47
|
+
name="predictive",
|
|
48
|
+
oauth_check=None,
|
|
49
|
+
directory="predictive",
|
|
50
|
+
package_prefix="datarobot_genai.drmcp.tools.predictive",
|
|
51
|
+
config_field_name="enable_predictive_tools",
|
|
52
|
+
),
|
|
53
|
+
ToolType.JIRA: ToolConfig(
|
|
54
|
+
name="jira",
|
|
55
|
+
oauth_check=lambda config: config.is_jira_oauth_configured,
|
|
56
|
+
directory="jira",
|
|
57
|
+
package_prefix="datarobot_genai.drmcp.tools.jira",
|
|
58
|
+
config_field_name="enable_jira_tools",
|
|
59
|
+
),
|
|
60
|
+
ToolType.CONFLUENCE: ToolConfig(
|
|
61
|
+
name="confluence",
|
|
62
|
+
oauth_check=lambda config: config.is_confluence_oauth_configured,
|
|
63
|
+
directory="confluence",
|
|
64
|
+
package_prefix="datarobot_genai.drmcp.tools.confluence",
|
|
65
|
+
config_field_name="enable_confluence_tools",
|
|
66
|
+
),
|
|
67
|
+
}
|
|
68
|
+
|
|
69
|
+
|
|
70
|
+
def get_tool_enable_config_name(tool_type: ToolType) -> str:
|
|
71
|
+
"""Get the configuration field name for enabling a tool."""
|
|
72
|
+
return TOOL_CONFIGS[tool_type]["config_field_name"]
|
|
73
|
+
|
|
74
|
+
|
|
75
|
+
def is_tool_enabled(tool_type: ToolType, config: "MCPServerConfig") -> bool:
|
|
76
|
+
"""
|
|
77
|
+
Check if a tool is enabled based on configuration.
|
|
78
|
+
|
|
79
|
+
Args:
|
|
80
|
+
tool_type: The type of tool to check
|
|
81
|
+
config: The server configuration
|
|
82
|
+
|
|
83
|
+
Returns
|
|
84
|
+
-------
|
|
85
|
+
True if the tool is enabled, False otherwise
|
|
86
|
+
"""
|
|
87
|
+
tool_config = TOOL_CONFIGS[tool_type]
|
|
88
|
+
enable_config_name = tool_config["config_field_name"]
|
|
89
|
+
is_enabled = getattr(config, enable_config_name)
|
|
90
|
+
|
|
91
|
+
# If tool is enabled, check OAuth requirements if needed
|
|
92
|
+
if is_enabled and tool_config["oauth_check"] is not None:
|
|
93
|
+
return tool_config["oauth_check"](config)
|
|
94
|
+
|
|
95
|
+
return is_enabled
|
|
@@ -16,6 +16,8 @@ import os
|
|
|
16
16
|
from collections.abc import AsyncGenerator
|
|
17
17
|
from contextlib import asynccontextmanager
|
|
18
18
|
|
|
19
|
+
import aiohttp
|
|
20
|
+
from aiohttp import ClientSession as HttpClientSession
|
|
19
21
|
from mcp import ClientSession
|
|
20
22
|
from mcp.client.streamable_http import streamablehttp_client
|
|
21
23
|
|
|
@@ -29,6 +31,11 @@ def get_dr_mcp_server_url() -> str | None:
|
|
|
29
31
|
return os.environ.get("DR_MCP_SERVER_URL")
|
|
30
32
|
|
|
31
33
|
|
|
34
|
+
def get_dr_mcp_server_http_url() -> str | None:
|
|
35
|
+
"""Get DataRobot MCP server http URL."""
|
|
36
|
+
return os.environ.get("DR_MCP_SERVER_HTTP_URL")
|
|
37
|
+
|
|
38
|
+
|
|
32
39
|
def get_openai_llm_client_config() -> dict[str, str]:
|
|
33
40
|
"""Get OpenAI LLM client configuration."""
|
|
34
41
|
openai_api_key = os.environ.get("OPENAI_API_KEY")
|
|
@@ -94,3 +101,25 @@ async def ete_test_mcp_session(
|
|
|
94
101
|
yield session
|
|
95
102
|
except asyncio.TimeoutError:
|
|
96
103
|
raise TimeoutError(f"Check if the MCP server is running at {get_dr_mcp_server_url()}")
|
|
104
|
+
|
|
105
|
+
|
|
106
|
+
@asynccontextmanager
|
|
107
|
+
async def ete_test_http_session(
|
|
108
|
+
additional_headers: dict[str, str] | None = None,
|
|
109
|
+
) -> AsyncGenerator[HttpClientSession, None]:
|
|
110
|
+
"""Create an HTTP session for each test that can connect to MCP custom http routes.
|
|
111
|
+
|
|
112
|
+
Parameters
|
|
113
|
+
----------
|
|
114
|
+
additional_headers : dict[str, str], optional
|
|
115
|
+
Additional headers to include in the HTTP session (e.g., auth headers for testing).
|
|
116
|
+
"""
|
|
117
|
+
headers = get_headers()
|
|
118
|
+
if additional_headers:
|
|
119
|
+
headers.update(additional_headers)
|
|
120
|
+
|
|
121
|
+
async with ete_test_mcp_session(additional_headers=additional_headers):
|
|
122
|
+
async with aiohttp.ClientSession(
|
|
123
|
+
base_url=get_dr_mcp_server_http_url(), headers=headers
|
|
124
|
+
) as client:
|
|
125
|
+
yield client
|