alita-sdk 0.3.351__py3-none-any.whl → 0.3.499__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (206) hide show
  1. alita_sdk/cli/__init__.py +10 -0
  2. alita_sdk/cli/__main__.py +17 -0
  3. alita_sdk/cli/agent/__init__.py +5 -0
  4. alita_sdk/cli/agent/default.py +258 -0
  5. alita_sdk/cli/agent_executor.py +155 -0
  6. alita_sdk/cli/agent_loader.py +215 -0
  7. alita_sdk/cli/agent_ui.py +228 -0
  8. alita_sdk/cli/agents.py +3601 -0
  9. alita_sdk/cli/callbacks.py +647 -0
  10. alita_sdk/cli/cli.py +168 -0
  11. alita_sdk/cli/config.py +306 -0
  12. alita_sdk/cli/context/__init__.py +30 -0
  13. alita_sdk/cli/context/cleanup.py +198 -0
  14. alita_sdk/cli/context/manager.py +731 -0
  15. alita_sdk/cli/context/message.py +285 -0
  16. alita_sdk/cli/context/strategies.py +289 -0
  17. alita_sdk/cli/context/token_estimation.py +127 -0
  18. alita_sdk/cli/formatting.py +182 -0
  19. alita_sdk/cli/input_handler.py +419 -0
  20. alita_sdk/cli/inventory.py +1256 -0
  21. alita_sdk/cli/mcp_loader.py +315 -0
  22. alita_sdk/cli/toolkit.py +327 -0
  23. alita_sdk/cli/toolkit_loader.py +85 -0
  24. alita_sdk/cli/tools/__init__.py +43 -0
  25. alita_sdk/cli/tools/approval.py +224 -0
  26. alita_sdk/cli/tools/filesystem.py +1751 -0
  27. alita_sdk/cli/tools/planning.py +389 -0
  28. alita_sdk/cli/tools/terminal.py +414 -0
  29. alita_sdk/community/__init__.py +64 -8
  30. alita_sdk/community/inventory/__init__.py +224 -0
  31. alita_sdk/community/inventory/config.py +257 -0
  32. alita_sdk/community/inventory/enrichment.py +2137 -0
  33. alita_sdk/community/inventory/extractors.py +1469 -0
  34. alita_sdk/community/inventory/ingestion.py +3172 -0
  35. alita_sdk/community/inventory/knowledge_graph.py +1457 -0
  36. alita_sdk/community/inventory/parsers/__init__.py +218 -0
  37. alita_sdk/community/inventory/parsers/base.py +295 -0
  38. alita_sdk/community/inventory/parsers/csharp_parser.py +907 -0
  39. alita_sdk/community/inventory/parsers/go_parser.py +851 -0
  40. alita_sdk/community/inventory/parsers/html_parser.py +389 -0
  41. alita_sdk/community/inventory/parsers/java_parser.py +593 -0
  42. alita_sdk/community/inventory/parsers/javascript_parser.py +629 -0
  43. alita_sdk/community/inventory/parsers/kotlin_parser.py +768 -0
  44. alita_sdk/community/inventory/parsers/markdown_parser.py +362 -0
  45. alita_sdk/community/inventory/parsers/python_parser.py +604 -0
  46. alita_sdk/community/inventory/parsers/rust_parser.py +858 -0
  47. alita_sdk/community/inventory/parsers/swift_parser.py +832 -0
  48. alita_sdk/community/inventory/parsers/text_parser.py +322 -0
  49. alita_sdk/community/inventory/parsers/yaml_parser.py +370 -0
  50. alita_sdk/community/inventory/patterns/__init__.py +61 -0
  51. alita_sdk/community/inventory/patterns/ast_adapter.py +380 -0
  52. alita_sdk/community/inventory/patterns/loader.py +348 -0
  53. alita_sdk/community/inventory/patterns/registry.py +198 -0
  54. alita_sdk/community/inventory/presets.py +535 -0
  55. alita_sdk/community/inventory/retrieval.py +1403 -0
  56. alita_sdk/community/inventory/toolkit.py +173 -0
  57. alita_sdk/community/inventory/visualize.py +1370 -0
  58. alita_sdk/configurations/bitbucket.py +94 -2
  59. alita_sdk/configurations/confluence.py +96 -1
  60. alita_sdk/configurations/gitlab.py +79 -0
  61. alita_sdk/configurations/jira.py +103 -0
  62. alita_sdk/configurations/testrail.py +88 -0
  63. alita_sdk/configurations/xray.py +93 -0
  64. alita_sdk/configurations/zephyr_enterprise.py +93 -0
  65. alita_sdk/configurations/zephyr_essential.py +75 -0
  66. alita_sdk/runtime/clients/artifact.py +1 -1
  67. alita_sdk/runtime/clients/client.py +214 -42
  68. alita_sdk/runtime/clients/mcp_discovery.py +342 -0
  69. alita_sdk/runtime/clients/mcp_manager.py +262 -0
  70. alita_sdk/runtime/clients/sandbox_client.py +373 -0
  71. alita_sdk/runtime/langchain/assistant.py +118 -30
  72. alita_sdk/runtime/langchain/constants.py +8 -1
  73. alita_sdk/runtime/langchain/document_loaders/AlitaDocxMammothLoader.py +315 -3
  74. alita_sdk/runtime/langchain/document_loaders/AlitaExcelLoader.py +103 -60
  75. alita_sdk/runtime/langchain/document_loaders/AlitaJSONLoader.py +4 -1
  76. alita_sdk/runtime/langchain/document_loaders/AlitaPowerPointLoader.py +41 -12
  77. alita_sdk/runtime/langchain/document_loaders/AlitaTableLoader.py +1 -1
  78. alita_sdk/runtime/langchain/document_loaders/constants.py +116 -99
  79. alita_sdk/runtime/langchain/interfaces/llm_processor.py +2 -2
  80. alita_sdk/runtime/langchain/langraph_agent.py +307 -71
  81. alita_sdk/runtime/langchain/utils.py +48 -8
  82. alita_sdk/runtime/llms/preloaded.py +2 -6
  83. alita_sdk/runtime/models/mcp_models.py +61 -0
  84. alita_sdk/runtime/toolkits/__init__.py +26 -0
  85. alita_sdk/runtime/toolkits/application.py +9 -2
  86. alita_sdk/runtime/toolkits/artifact.py +18 -6
  87. alita_sdk/runtime/toolkits/datasource.py +13 -6
  88. alita_sdk/runtime/toolkits/mcp.py +780 -0
  89. alita_sdk/runtime/toolkits/planning.py +178 -0
  90. alita_sdk/runtime/toolkits/tools.py +205 -55
  91. alita_sdk/runtime/toolkits/vectorstore.py +9 -4
  92. alita_sdk/runtime/tools/__init__.py +11 -3
  93. alita_sdk/runtime/tools/application.py +7 -0
  94. alita_sdk/runtime/tools/artifact.py +225 -12
  95. alita_sdk/runtime/tools/function.py +95 -5
  96. alita_sdk/runtime/tools/graph.py +10 -4
  97. alita_sdk/runtime/tools/image_generation.py +212 -0
  98. alita_sdk/runtime/tools/llm.py +494 -102
  99. alita_sdk/runtime/tools/mcp_inspect_tool.py +284 -0
  100. alita_sdk/runtime/tools/mcp_remote_tool.py +181 -0
  101. alita_sdk/runtime/tools/mcp_server_tool.py +4 -4
  102. alita_sdk/runtime/tools/planning/__init__.py +36 -0
  103. alita_sdk/runtime/tools/planning/models.py +246 -0
  104. alita_sdk/runtime/tools/planning/wrapper.py +607 -0
  105. alita_sdk/runtime/tools/router.py +2 -1
  106. alita_sdk/runtime/tools/sandbox.py +180 -79
  107. alita_sdk/runtime/tools/vectorstore.py +22 -21
  108. alita_sdk/runtime/tools/vectorstore_base.py +125 -52
  109. alita_sdk/runtime/utils/AlitaCallback.py +106 -20
  110. alita_sdk/runtime/utils/mcp_client.py +465 -0
  111. alita_sdk/runtime/utils/mcp_oauth.py +244 -0
  112. alita_sdk/runtime/utils/mcp_sse_client.py +405 -0
  113. alita_sdk/runtime/utils/mcp_tools_discovery.py +124 -0
  114. alita_sdk/runtime/utils/streamlit.py +40 -13
  115. alita_sdk/runtime/utils/toolkit_utils.py +28 -9
  116. alita_sdk/runtime/utils/utils.py +12 -0
  117. alita_sdk/tools/__init__.py +77 -33
  118. alita_sdk/tools/ado/repos/__init__.py +7 -6
  119. alita_sdk/tools/ado/repos/repos_wrapper.py +11 -11
  120. alita_sdk/tools/ado/test_plan/__init__.py +7 -7
  121. alita_sdk/tools/ado/wiki/__init__.py +7 -11
  122. alita_sdk/tools/ado/wiki/ado_wrapper.py +89 -15
  123. alita_sdk/tools/ado/work_item/__init__.py +7 -11
  124. alita_sdk/tools/ado/work_item/ado_wrapper.py +17 -8
  125. alita_sdk/tools/advanced_jira_mining/__init__.py +8 -7
  126. alita_sdk/tools/aws/delta_lake/__init__.py +11 -9
  127. alita_sdk/tools/azure_ai/search/__init__.py +7 -6
  128. alita_sdk/tools/base_indexer_toolkit.py +345 -70
  129. alita_sdk/tools/bitbucket/__init__.py +9 -8
  130. alita_sdk/tools/bitbucket/api_wrapper.py +50 -6
  131. alita_sdk/tools/browser/__init__.py +4 -4
  132. alita_sdk/tools/carrier/__init__.py +4 -6
  133. alita_sdk/tools/chunkers/__init__.py +3 -1
  134. alita_sdk/tools/chunkers/sematic/json_chunker.py +1 -0
  135. alita_sdk/tools/chunkers/sematic/markdown_chunker.py +97 -6
  136. alita_sdk/tools/chunkers/sematic/proposal_chunker.py +1 -1
  137. alita_sdk/tools/chunkers/universal_chunker.py +270 -0
  138. alita_sdk/tools/cloud/aws/__init__.py +7 -6
  139. alita_sdk/tools/cloud/azure/__init__.py +7 -6
  140. alita_sdk/tools/cloud/gcp/__init__.py +7 -6
  141. alita_sdk/tools/cloud/k8s/__init__.py +7 -6
  142. alita_sdk/tools/code/linter/__init__.py +7 -7
  143. alita_sdk/tools/code/loaders/codesearcher.py +3 -2
  144. alita_sdk/tools/code/sonar/__init__.py +8 -7
  145. alita_sdk/tools/code_indexer_toolkit.py +199 -0
  146. alita_sdk/tools/confluence/__init__.py +9 -8
  147. alita_sdk/tools/confluence/api_wrapper.py +171 -75
  148. alita_sdk/tools/confluence/loader.py +10 -0
  149. alita_sdk/tools/custom_open_api/__init__.py +9 -4
  150. alita_sdk/tools/elastic/__init__.py +8 -7
  151. alita_sdk/tools/elitea_base.py +492 -52
  152. alita_sdk/tools/figma/__init__.py +7 -7
  153. alita_sdk/tools/figma/api_wrapper.py +2 -1
  154. alita_sdk/tools/github/__init__.py +9 -9
  155. alita_sdk/tools/github/api_wrapper.py +9 -26
  156. alita_sdk/tools/github/github_client.py +62 -2
  157. alita_sdk/tools/gitlab/__init__.py +8 -8
  158. alita_sdk/tools/gitlab/api_wrapper.py +135 -33
  159. alita_sdk/tools/gitlab_org/__init__.py +7 -8
  160. alita_sdk/tools/google/bigquery/__init__.py +11 -12
  161. alita_sdk/tools/google_places/__init__.py +8 -7
  162. alita_sdk/tools/jira/__init__.py +9 -7
  163. alita_sdk/tools/jira/api_wrapper.py +100 -52
  164. alita_sdk/tools/keycloak/__init__.py +8 -7
  165. alita_sdk/tools/localgit/local_git.py +56 -54
  166. alita_sdk/tools/memory/__init__.py +1 -1
  167. alita_sdk/tools/non_code_indexer_toolkit.py +3 -2
  168. alita_sdk/tools/ocr/__init__.py +8 -7
  169. alita_sdk/tools/openapi/__init__.py +10 -1
  170. alita_sdk/tools/pandas/__init__.py +8 -7
  171. alita_sdk/tools/postman/__init__.py +7 -8
  172. alita_sdk/tools/postman/api_wrapper.py +19 -8
  173. alita_sdk/tools/postman/postman_analysis.py +8 -1
  174. alita_sdk/tools/pptx/__init__.py +8 -9
  175. alita_sdk/tools/qtest/__init__.py +16 -11
  176. alita_sdk/tools/qtest/api_wrapper.py +1784 -88
  177. alita_sdk/tools/rally/__init__.py +7 -8
  178. alita_sdk/tools/report_portal/__init__.py +9 -7
  179. alita_sdk/tools/salesforce/__init__.py +7 -7
  180. alita_sdk/tools/servicenow/__init__.py +10 -10
  181. alita_sdk/tools/sharepoint/__init__.py +7 -6
  182. alita_sdk/tools/sharepoint/api_wrapper.py +127 -36
  183. alita_sdk/tools/sharepoint/authorization_helper.py +191 -1
  184. alita_sdk/tools/sharepoint/utils.py +8 -2
  185. alita_sdk/tools/slack/__init__.py +7 -6
  186. alita_sdk/tools/sql/__init__.py +8 -7
  187. alita_sdk/tools/sql/api_wrapper.py +71 -23
  188. alita_sdk/tools/testio/__init__.py +7 -6
  189. alita_sdk/tools/testrail/__init__.py +8 -9
  190. alita_sdk/tools/utils/__init__.py +26 -4
  191. alita_sdk/tools/utils/content_parser.py +88 -60
  192. alita_sdk/tools/utils/text_operations.py +254 -0
  193. alita_sdk/tools/vector_adapters/VectorStoreAdapter.py +76 -26
  194. alita_sdk/tools/xray/__init__.py +9 -7
  195. alita_sdk/tools/zephyr/__init__.py +7 -6
  196. alita_sdk/tools/zephyr_enterprise/__init__.py +8 -6
  197. alita_sdk/tools/zephyr_essential/__init__.py +7 -6
  198. alita_sdk/tools/zephyr_essential/api_wrapper.py +12 -13
  199. alita_sdk/tools/zephyr_scale/__init__.py +7 -6
  200. alita_sdk/tools/zephyr_squad/__init__.py +7 -6
  201. {alita_sdk-0.3.351.dist-info → alita_sdk-0.3.499.dist-info}/METADATA +147 -2
  202. {alita_sdk-0.3.351.dist-info → alita_sdk-0.3.499.dist-info}/RECORD +206 -130
  203. alita_sdk-0.3.499.dist-info/entry_points.txt +2 -0
  204. {alita_sdk-0.3.351.dist-info → alita_sdk-0.3.499.dist-info}/WHEEL +0 -0
  205. {alita_sdk-0.3.351.dist-info → alita_sdk-0.3.499.dist-info}/licenses/LICENSE +0 -0
  206. {alita_sdk-0.3.351.dist-info → alita_sdk-0.3.499.dist-info}/top_level.txt +0 -0
@@ -1,4 +1,6 @@
1
1
  import hashlib
2
+ import io
3
+ import json
2
4
  import logging
3
5
  import re
4
6
  from typing import Any, Optional, Generator, List
@@ -6,10 +8,12 @@ from typing import Any, Optional, Generator, List
6
8
  from langchain_core.callbacks import dispatch_custom_event
7
9
  from langchain_core.documents import Document
8
10
  from langchain_core.tools import ToolException
11
+ from openpyxl.workbook.workbook import Workbook
9
12
  from pydantic import create_model, Field, model_validator
10
13
 
11
14
  from ...tools.non_code_indexer_toolkit import NonCodeIndexerToolkit
12
15
  from ...tools.utils.available_tools_decorator import extend_with_parent_available_tools
16
+ from ...tools.elitea_base import extend_with_file_operations
13
17
  from ...runtime.utils.utils import IndexerKeywords
14
18
 
15
19
 
@@ -28,10 +32,36 @@ class ArtifactWrapper(NonCodeIndexerToolkit):
28
32
  return super().validate_toolkit(values)
29
33
 
30
34
  def list_files(self, bucket_name = None, return_as_string = True):
31
- return self.artifact.list(bucket_name, return_as_string)
35
+ """List all files in the artifact bucket with API download links."""
36
+ result = self.artifact.list(bucket_name, return_as_string=False)
37
+
38
+ # Add API download link to each file
39
+ if isinstance(result, dict) and 'rows' in result:
40
+ bucket = bucket_name or self.bucket
41
+
42
+ # Get base_url and project_id from alita client
43
+ base_url = getattr(self.alita, 'base_url', '').rstrip('/')
44
+ project_id = getattr(self.alita, 'project_id', '')
45
+
46
+ for file_info in result['rows']:
47
+ if 'name' in file_info:
48
+ # Generate API download link
49
+ file_name = file_info['name']
50
+ file_info['link'] = f"{base_url}/api/v2/artifacts/artifact/default/{project_id}/{bucket}/{file_name}"
51
+
52
+ return str(result) if return_as_string else result
32
53
 
33
54
  def create_file(self, filename: str, filedata: str, bucket_name = None):
34
- result = self.artifact.create(filename, filedata, bucket_name)
55
+ # Sanitize filename to prevent regex errors during indexing
56
+ sanitized_filename, was_modified = self._sanitize_filename(filename)
57
+ if was_modified:
58
+ logging.warning(f"Filename sanitized: '{filename}' -> '{sanitized_filename}'")
59
+
60
+ if sanitized_filename.endswith(".xlsx"):
61
+ data = json.loads(filedata)
62
+ filedata = self.create_xlsx_filedata(data)
63
+
64
+ result = self.artifact.create(sanitized_filename, filedata, bucket_name)
35
65
 
36
66
  # Dispatch custom event for file creation
37
67
  dispatch_custom_event("file_modified", {
@@ -44,8 +74,63 @@ class ArtifactWrapper(NonCodeIndexerToolkit):
44
74
  "bucket": bucket_name or self.bucket
45
75
  }
46
76
  })
47
-
77
+
48
78
  return result
79
+
80
+ @staticmethod
81
+ def _sanitize_filename(filename: str) -> tuple:
82
+ """Sanitize filename for safe storage and regex pattern matching."""
83
+ from pathlib import Path
84
+
85
+ if not filename or not filename.strip():
86
+ return "unnamed_file", True
87
+
88
+ original = filename
89
+ path_obj = Path(filename)
90
+ name = path_obj.stem
91
+ extension = path_obj.suffix
92
+
93
+ # Whitelist: alphanumeric, underscore, hyphen, space, Unicode letters/digits
94
+ sanitized_name = re.sub(r'[^\w\s-]', '', name, flags=re.UNICODE)
95
+ sanitized_name = re.sub(r'[-\s]+', '-', sanitized_name)
96
+ sanitized_name = sanitized_name.strip('-').strip()
97
+
98
+ if not sanitized_name:
99
+ sanitized_name = "file"
100
+
101
+ if extension:
102
+ extension = re.sub(r'[^\w.-]', '', extension, flags=re.UNICODE)
103
+
104
+ sanitized = sanitized_name + extension
105
+ return sanitized, (sanitized != original)
106
+
107
+ def create_xlsx_filedata(self, data: dict[str, list[list]]) -> bytes:
108
+ try:
109
+ workbook = Workbook()
110
+
111
+ first_sheet = True
112
+ for sheet_name, sheet_data in data.items():
113
+ if first_sheet:
114
+ sheet = workbook.active
115
+ sheet.title = sheet_name
116
+ first_sheet = False
117
+ else:
118
+ sheet = workbook.create_sheet(title=sheet_name)
119
+
120
+ for row in sheet_data:
121
+ sheet.append(row)
122
+
123
+ file_buffer = io.BytesIO()
124
+ workbook.save(file_buffer)
125
+ file_buffer.seek(0)
126
+
127
+ return file_buffer.read()
128
+
129
+ except json.JSONDecodeError:
130
+ raise ValueError("Invalid JSON format for .xlsx file data.")
131
+ except Exception as e:
132
+ raise ValueError(f"Error processing .xlsx file data: {e}")
133
+
49
134
 
50
135
  def read_file(self,
51
136
  filename: str,
@@ -61,6 +146,94 @@ class ArtifactWrapper(NonCodeIndexerToolkit):
61
146
  sheet_name=sheet_name,
62
147
  excel_by_sheets=excel_by_sheets,
63
148
  llm=self.llm)
149
+
150
+ def _read_file(
151
+ self,
152
+ file_path: str,
153
+ branch: str = None,
154
+ bucket_name: str = None,
155
+ **kwargs
156
+ ) -> str:
157
+ """
158
+ Read a file from artifact bucket with optional partial read support.
159
+
160
+ Parameters:
161
+ file_path: Name of the file in the bucket
162
+ branch: Not used for artifacts (kept for API consistency)
163
+ bucket_name: Name of the bucket (uses default if None)
164
+ **kwargs: Additional parameters (offset, limit, head, tail) - currently ignored,
165
+ partial read handled client-side by base class methods
166
+
167
+ Returns:
168
+ File content as string
169
+ """
170
+ return self.read_file(filename=file_path, bucket_name=bucket_name)
171
+
172
+ def _write_file(
173
+ self,
174
+ file_path: str,
175
+ content: str,
176
+ branch: str = None,
177
+ commit_message: str = None,
178
+ bucket_name: str = None
179
+ ) -> str:
180
+ """
181
+ Write content to a file (create or overwrite).
182
+
183
+ Parameters:
184
+ file_path: Name of the file in the bucket
185
+ content: New file content
186
+ branch: Not used for artifacts (kept for API consistency)
187
+ commit_message: Not used for artifacts (kept for API consistency)
188
+ bucket_name: Name of the bucket (uses default if None)
189
+
190
+ Returns:
191
+ Success message
192
+ """
193
+ try:
194
+ # Sanitize filename
195
+ sanitized_filename, was_modified = self._sanitize_filename(file_path)
196
+ if was_modified:
197
+ logging.warning(f"Filename sanitized: '{file_path}' -> '{sanitized_filename}'")
198
+
199
+ # Check if file exists
200
+ try:
201
+ self.artifact.get(artifact_name=sanitized_filename, bucket_name=bucket_name, llm=self.llm)
202
+ # File exists, overwrite it
203
+ result = self.artifact.overwrite(sanitized_filename, content, bucket_name)
204
+
205
+ # Dispatch custom event
206
+ dispatch_custom_event("file_modified", {
207
+ "message": f"File '{sanitized_filename}' updated successfully",
208
+ "filename": sanitized_filename,
209
+ "tool_name": "edit_file",
210
+ "toolkit": "artifact",
211
+ "operation_type": "modify",
212
+ "meta": {
213
+ "bucket": bucket_name or self.bucket
214
+ }
215
+ })
216
+
217
+ return f"Updated file {sanitized_filename}"
218
+ except:
219
+ # File doesn't exist, create it
220
+ result = self.artifact.create(sanitized_filename, content, bucket_name)
221
+
222
+ # Dispatch custom event
223
+ dispatch_custom_event("file_modified", {
224
+ "message": f"File '{sanitized_filename}' created successfully",
225
+ "filename": sanitized_filename,
226
+ "tool_name": "edit_file",
227
+ "toolkit": "artifact",
228
+ "operation_type": "create",
229
+ "meta": {
230
+ "bucket": bucket_name or self.bucket
231
+ }
232
+ })
233
+
234
+ return f"Created file {sanitized_filename}"
235
+ except Exception as e:
236
+ raise ToolException(f"Unable to write file {file_path}: {str(e)}")
64
237
 
65
238
  def delete_file(self, filename: str, bucket_name = None):
66
239
  return self.artifact.delete(filename, bucket_name)
@@ -100,7 +273,11 @@ class ArtifactWrapper(NonCodeIndexerToolkit):
100
273
  return result
101
274
 
102
275
  def create_new_bucket(self, bucket_name: str, expiration_measure = "weeks", expiration_value = 1):
103
- return self.artifact.client.create_bucket(bucket_name, expiration_measure, expiration_value)
276
+ # Sanitize bucket name: replace underscores with hyphens and ensure lowercase
277
+ sanitized_name = bucket_name.replace('_', '-').lower()
278
+ if sanitized_name != bucket_name:
279
+ logging.warning(f"Bucket name '{bucket_name}' was sanitized to '{sanitized_name}' (underscores replaced with hyphens, converted to lowercase)")
280
+ return self.artifact.client.create_bucket(sanitized_name, expiration_measure, expiration_value)
104
281
 
105
282
  def _index_tool_params(self):
106
283
  return {
@@ -135,13 +312,13 @@ class ArtifactWrapper(NonCodeIndexerToolkit):
135
312
  file_name = file['name']
136
313
 
137
314
  # Check if file should be skipped based on skip_extensions
138
- if any(re.match(pattern.replace('*', '.*') + '$', file_name, re.IGNORECASE)
315
+ if any(re.match(re.escape(pattern).replace(r'\*', '.*') + '$', file_name, re.IGNORECASE)
139
316
  for pattern in skip_extensions):
140
317
  continue
141
318
 
142
319
  # Check if file should be included based on include_extensions
143
320
  # If include_extensions is empty, process all files (that weren't skipped)
144
- if include_extensions and not (any(re.match(pattern.replace('*', '.*') + '$', file_name, re.IGNORECASE)
321
+ if include_extensions and not (any(re.match(re.escape(pattern).replace(r'\*', '.*') + '$', file_name, re.IGNORECASE)
145
322
  for pattern in include_extensions)):
146
323
  continue
147
324
 
@@ -166,17 +343,20 @@ class ArtifactWrapper(NonCodeIndexerToolkit):
166
343
  document.metadata[IndexerKeywords.CONTENT_FILE_NAME.value] = document.metadata['name']
167
344
  yield document
168
345
  except Exception as e:
169
- logging.error(f"Failed while parsing the file '{document.metadata['name']}': {e}")
346
+ logger.error(f"Failed while parsing the file '{document.metadata['name']}': {e}")
170
347
  yield document
171
348
 
172
- @extend_with_parent_available_tools
349
+ @extend_with_file_operations
173
350
  def get_available_tools(self):
351
+ """Get available tools, including indexing tools only if vector store is configured."""
174
352
  bucket_name = (Optional[str], Field(description="Name of the bucket to work with."
175
353
  "If bucket is not specified by user directly, the name should be taken from chat history."
176
354
  "If bucket never mentioned in chat, the name will be taken from tool configuration."
177
355
  " ***IMPORTANT*** Underscore `_` is prohibited in bucket name and should be replaced by `-`",
178
356
  default=None))
179
- return [
357
+
358
+ # Basic artifact tools (always available)
359
+ basic_tools = [
180
360
  {
181
361
  "ref": self.list_files,
182
362
  "name": "listFiles",
@@ -190,7 +370,17 @@ class ArtifactWrapper(NonCodeIndexerToolkit):
190
370
  "args_schema": create_model(
191
371
  "createFile",
192
372
  filename=(str, Field(description="Filename")),
193
- filedata=(str, Field(description="Stringified content of the file")),
373
+ filedata=(str, Field(description="""Stringified content of the file.
374
+ Example for .xlsx filedata format:
375
+ {
376
+ "Sheet1":[
377
+ ["Name", "Age", "City"],
378
+ ["Alice", 25, "New York"],
379
+ ["Bob", 30, "San Francisco"],
380
+ ["Charlie", 35, "Los Angeles"]
381
+ ]
382
+ }
383
+ """)),
194
384
  bucket_name=bucket_name
195
385
  )
196
386
  },
@@ -251,11 +441,34 @@ class ArtifactWrapper(NonCodeIndexerToolkit):
251
441
  "description": "Creates new bucket specified by user.",
252
442
  "args_schema": create_model(
253
443
  "createNewBucket",
254
- bucket_name=(str, Field(description="Bucket name to create. ***IMPORTANT*** Underscore `_` is prohibited in bucket name and should be replaced by `-`.")),
444
+ bucket_name=(str, Field(
445
+ description="Bucket name to create. Must start with lowercase letter and contain only lowercase letters, numbers, and hyphens. Underscores will be automatically converted to hyphens.",
446
+ pattern=r'^[a-z][a-z0-9_-]*$' # Allow underscores in input, will be sanitized
447
+ )),
255
448
  expiration_measure=(Optional[str], Field(description="Measure of expiration time for bucket configuration."
256
449
  "Possible values: `days`, `weeks`, `months`, `years`.",
257
450
  default="weeks")),
258
451
  expiration_value=(Optional[int], Field(description="Expiration time values.", default=1))
259
452
  )
260
453
  }
261
- ]
454
+ ]
455
+
456
+ # Add indexing tools only if vector store is configured
457
+ has_vector_config = (
458
+ hasattr(self, 'embedding_model') and self.embedding_model and
459
+ hasattr(self, 'pgvector_configuration') and self.pgvector_configuration
460
+ )
461
+
462
+ if has_vector_config:
463
+ try:
464
+ # Get indexing tools from parent class
465
+ indexing_tools = super(ArtifactWrapper, self).get_available_tools()
466
+ return indexing_tools + basic_tools
467
+ except Exception as e:
468
+ # If getting parent tools fails, log warning and return basic tools only
469
+ logging.warning(f"Failed to load indexing tools: {e}. Only basic artifact tools will be available.")
470
+ return basic_tools
471
+ else:
472
+ # No vector store config, return basic tools only
473
+ logging.info("Vector store not configured. Indexing tools (index_data, search_index, etc.) are not available.")
474
+ return basic_tools
@@ -1,18 +1,33 @@
1
+ import json
1
2
  import logging
3
+ from copy import deepcopy
2
4
  from json import dumps
3
5
 
4
6
  from langchain_core.callbacks import dispatch_custom_event
5
7
  from langchain_core.messages import ToolCall
6
8
  from langchain_core.runnables import RunnableConfig
7
9
  from langchain_core.tools import BaseTool, ToolException
8
- from typing import Any, Optional, Union, Annotated
10
+ from typing import Any, Optional, Union
9
11
  from langchain_core.utils.function_calling import convert_to_openai_tool
10
12
  from pydantic import ValidationError
13
+
11
14
  from ..langchain.utils import propagate_the_input_mapping
12
15
 
13
16
  logger = logging.getLogger(__name__)
14
17
 
15
18
 
19
+ def replace_escaped_newlines(data):
20
+ """
21
+ Replace \\n with \n in all string values recursively.
22
+ Required for sanitization of state variables in code node
23
+ """
24
+ if isinstance(data, dict):
25
+ return {key: replace_escaped_newlines(value) for key, value in data.items()}
26
+ elif isinstance(data, str):
27
+ return data.replace('\\n', '\n')
28
+ else:
29
+ return data
30
+
16
31
  class FunctionTool(BaseTool):
17
32
  name: str = 'FunctionalTool'
18
33
  description: str = 'This is direct call node for tools'
@@ -21,6 +36,61 @@ class FunctionTool(BaseTool):
21
36
  input_variables: Optional[list[str]] = None
22
37
  input_mapping: Optional[dict[str, dict]] = None
23
38
  output_variables: Optional[list[str]] = None
39
+ structured_output: Optional[bool] = False
40
+ alita_client: Optional[Any] = None
41
+
42
+ def _prepare_pyodide_input(self, state: Union[str, dict, ToolCall]) -> str:
43
+ """Prepare input for PyodideSandboxTool by injecting state into the code block."""
44
+ # add state into the code block here since it might be changed during the execution of the code
45
+ state_copy = replace_escaped_newlines(deepcopy(state))
46
+
47
+ del state_copy['messages'] # remove messages to avoid issues with pickling without langchain-core
48
+ # inject state into the code block as alita_state variable
49
+ state_json = json.dumps(state_copy, ensure_ascii=False)
50
+ pyodide_predata = f'#state dict\nimport json\nalita_state = json.loads({json.dumps(state_json)})\n'
51
+
52
+ return pyodide_predata
53
+
54
+ def _handle_pyodide_output(self, tool_result: Any) -> dict:
55
+ """Handle output processing for PyodideSandboxTool results."""
56
+ tool_result_converted = {}
57
+
58
+ if self.output_variables:
59
+ for var in self.output_variables:
60
+ if var == "messages":
61
+ tool_result_converted.update(
62
+ {"messages": [{"role": "assistant", "content": dumps(tool_result)}]})
63
+ continue
64
+ if isinstance(tool_result, dict) and var in tool_result:
65
+ tool_result_converted[var] = tool_result[var]
66
+ else:
67
+ # handler in case user points to a var that is not in the output of the tool
68
+ tool_result_converted[var] = tool_result.get('result',
69
+ tool_result.get('error') if tool_result.get('error')
70
+ else 'Execution result is missing')
71
+ else:
72
+ tool_result_converted.update({"messages": [{"role": "assistant", "content": dumps(tool_result)}]})
73
+
74
+ if self.structured_output:
75
+ # execute code tool and update state variables
76
+ try:
77
+ result_value = tool_result.get('result', {})
78
+ if isinstance(result_value, dict):
79
+ tool_result_converted.update(result_value)
80
+ elif isinstance(result_value, list):
81
+ # Handle list case - could wrap in a key or handle differently based on requirements
82
+ tool_result_converted.update({"result": result_value})
83
+ else:
84
+ # Handle JSON string case
85
+ tool_result_converted.update(json.loads(result_value))
86
+ except json.JSONDecodeError:
87
+ logger.error(f"JSONDecodeError: {tool_result}")
88
+
89
+ return tool_result_converted
90
+
91
+ def _is_pyodide_tool(self) -> bool:
92
+ """Check if the current tool is a PyodideSandboxTool."""
93
+ return self.tool.name.lower() == 'pyodide_sandbox'
24
94
 
25
95
  def invoke(
26
96
  self,
@@ -31,8 +101,15 @@ class FunctionTool(BaseTool):
31
101
  params = convert_to_openai_tool(self.tool).get(
32
102
  'function', {'parameters': {}}).get(
33
103
  'parameters', {'properties': {}}).get('properties', {})
104
+
34
105
  func_args = propagate_the_input_mapping(input_mapping=self.input_mapping, input_variables=self.input_variables,
35
106
  state=state)
107
+
108
+ # special handler for PyodideSandboxTool
109
+ if self._is_pyodide_tool():
110
+ # replace new lines in strings in code block
111
+ code = func_args['code'].replace('\\n', '\\\\n')
112
+ func_args['code'] = f"{self._prepare_pyodide_input(state)}\n{code}"
36
113
  try:
37
114
  tool_result = self.tool.invoke(func_args, config, **kwargs)
38
115
  dispatch_custom_event(
@@ -44,17 +121,30 @@ class FunctionTool(BaseTool):
44
121
  }, config=config
45
122
  )
46
123
  logger.info(f"ToolNode response: {tool_result}")
124
+
125
+ # handler for PyodideSandboxTool
126
+ if self._is_pyodide_tool():
127
+ return self._handle_pyodide_output(tool_result)
128
+
47
129
  if not self.output_variables:
48
130
  return {"messages": [{"role": "assistant", "content": dumps(tool_result)}]}
49
131
  else:
50
- if self.output_variables[0] == "messages":
51
- return {
132
+ if "messages" in self.output_variables:
133
+ messages_dict = {
52
134
  "messages": [{
53
135
  "role": "assistant",
54
- "content": dumps(tool_result) if not isinstance(tool_result, ToolException) else str(
55
- tool_result)
136
+ "content": dumps(tool_result)
137
+ if not isinstance(tool_result, ToolException) and not isinstance(tool_result, str)
138
+ else str(tool_result)
56
139
  }]
57
140
  }
141
+ for var in self.output_variables:
142
+ if var != "messages":
143
+ if isinstance(tool_result, dict) and var in tool_result:
144
+ messages_dict[var] = tool_result[var]
145
+ else:
146
+ messages_dict[var] = tool_result
147
+ return messages_dict
58
148
  else:
59
149
  return { self.output_variables[0]: tool_result }
60
150
  except ValidationError:
@@ -47,8 +47,8 @@ def formulate_query(kwargs):
47
47
 
48
48
 
49
49
  class GraphTool(BaseTool):
50
- name: str
51
- description: str
50
+ name: str = 'GraphTool'
51
+ description: str = 'Graph tool for tools'
52
52
  graph: CompiledStateGraph
53
53
  args_schema: Type[BaseModel] = graphToolSchema
54
54
  return_type: str = "str"
@@ -65,10 +65,16 @@ class GraphTool(BaseTool):
65
65
  all_kwargs = {**kwargs, **extras, **schema_values}
66
66
  if config is None:
67
67
  config = {}
68
- return self._run(*config, **all_kwargs)
68
+ # Pass the config to the _run empty or the one passed from the parent executor.
69
+ return self._run(config, **all_kwargs)
69
70
 
70
71
  def _run(self, *args, **kwargs):
71
- response = self.graph.invoke(formulate_query(kwargs))
72
+ config = None
73
+ # From invoke method we are passing only 1 arg so it is safe to do this condition and config assignment.
74
+ # Default to None is safe because it will be checked also on the langchain side.
75
+ if args:
76
+ config = args[0]
77
+ response = self.graph.invoke(formulate_query(kwargs), config=config)
72
78
  if self.return_type == "str":
73
79
  return response["output"]
74
80
  else: