alita-sdk 0.3.206__py3-none-any.whl → 0.3.207__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- alita_sdk/runtime/clients/client.py +269 -6
- alita_sdk/runtime/langchain/langraph_agent.py +6 -1
- alita_sdk/runtime/langchain/store_manager.py +4 -4
- alita_sdk/runtime/toolkits/tools.py +11 -20
- alita_sdk/runtime/utils/streamlit.py +472 -192
- alita_sdk/runtime/utils/toolkit_runtime.py +147 -0
- alita_sdk/runtime/utils/toolkit_utils.py +157 -0
- alita_sdk/tools/memory/__init__.py +54 -10
- {alita_sdk-0.3.206.dist-info → alita_sdk-0.3.207.dist-info}/METADATA +1 -1
- {alita_sdk-0.3.206.dist-info → alita_sdk-0.3.207.dist-info}/RECORD +13 -20
- alita_sdk/community/analysis/__init__.py +0 -0
- alita_sdk/community/analysis/ado_analyse/__init__.py +0 -103
- alita_sdk/community/analysis/ado_analyse/api_wrapper.py +0 -261
- alita_sdk/community/analysis/github_analyse/__init__.py +0 -98
- alita_sdk/community/analysis/github_analyse/api_wrapper.py +0 -166
- alita_sdk/community/analysis/gitlab_analyse/__init__.py +0 -110
- alita_sdk/community/analysis/gitlab_analyse/api_wrapper.py +0 -172
- alita_sdk/community/analysis/jira_analyse/__init__.py +0 -141
- alita_sdk/community/analysis/jira_analyse/api_wrapper.py +0 -252
- {alita_sdk-0.3.206.dist-info → alita_sdk-0.3.207.dist-info}/WHEEL +0 -0
- {alita_sdk-0.3.206.dist-info → alita_sdk-0.3.207.dist-info}/licenses/LICENSE +0 -0
- {alita_sdk-0.3.206.dist-info → alita_sdk-0.3.207.dist-info}/top_level.txt +0 -0
@@ -1,261 +0,0 @@
|
|
1
|
-
import logging
|
2
|
-
|
3
|
-
from typing import Optional, Dict, Any
|
4
|
-
from langchain_core.callbacks import dispatch_custom_event
|
5
|
-
from pydantic import BaseModel, Field
|
6
|
-
|
7
|
-
from elitea_analyse.utils.constants import OUTPUT_WORK_ITEMS_FILE
|
8
|
-
from elitea_analyse.ado.azure_search import AzureSearch
|
9
|
-
from elitea_analyse.ado.main import (
|
10
|
-
OUTPUT_WORK_ITEMS,
|
11
|
-
get_work_items_several_projects,
|
12
|
-
get_commits_several_projects,
|
13
|
-
get_merge_requests_several_projects,
|
14
|
-
get_pipelines_runs_several_projects,
|
15
|
-
)
|
16
|
-
|
17
|
-
|
18
|
-
from alita_sdk.tools.elitea_base import BaseToolApiWrapper
|
19
|
-
|
20
|
-
from alita_sdk.runtime.utils.save_dataframe import save_dataframe_to_artifact
|
21
|
-
from alita_sdk.runtime.tools.artifact import ArtifactWrapper
|
22
|
-
from alita_sdk.runtime.utils.logging import with_streamlit_logs
|
23
|
-
|
24
|
-
|
25
|
-
logger = logging.getLogger(__name__)
|
26
|
-
|
27
|
-
class NoInputArgs(BaseModel):
|
28
|
-
pass
|
29
|
-
|
30
|
-
class GetAdoWorkItemsArgs(BaseModel):
|
31
|
-
resolved_after: str = Field(description="Resolveed after date (i.e. 2023-01-01)")
|
32
|
-
updated_after: str = Field(description="Updated after date (i.e. 2023-01-01)")
|
33
|
-
created_after: str = Field(description="Created after date (i.e. 2023-01-01)")
|
34
|
-
area: Optional[str] = Field(description="Area path filter.", default="")
|
35
|
-
project_keys: Optional[str] = Field(
|
36
|
-
description="One or more projects keys separated with comma.", default=""
|
37
|
-
)
|
38
|
-
|
39
|
-
|
40
|
-
class AdoCommitsArgs(BaseModel):
|
41
|
-
project_keys: Optional[str] = Field(
|
42
|
-
description="One or more projects keys separated with comma.", default=""
|
43
|
-
)
|
44
|
-
since_date: str = Field(description="Get commits after this date 'YYYY-MM-DD'")
|
45
|
-
|
46
|
-
|
47
|
-
class AdoPipelinesArgs(BaseModel):
|
48
|
-
project_keys: Optional[str] = Field(
|
49
|
-
description="One or more projects keys separated with comma.", default=""
|
50
|
-
)
|
51
|
-
|
52
|
-
|
53
|
-
class AdoAnalyseWrapper(BaseToolApiWrapper):
|
54
|
-
artifacts_wrapper: ArtifactWrapper
|
55
|
-
project_keys: str # Comma-separated list of Azure DevOps project names
|
56
|
-
default_branch_name: str = "main"
|
57
|
-
area: str = ""
|
58
|
-
ado_search: AzureSearch # Azure DevOps search client
|
59
|
-
|
60
|
-
class Config:
|
61
|
-
arbitrary_types_allowed = True
|
62
|
-
|
63
|
-
def get_projects_list(self):
|
64
|
-
"""
|
65
|
-
Get all projects in the organization that the authenticated user has access to.
|
66
|
-
Details on a page: https://docs.microsoft.com/en-us/rest/api/azure/devops/core/projects/list
|
67
|
-
"""
|
68
|
-
result = self.ado_search.get_projects_list()
|
69
|
-
|
70
|
-
save_dataframe_to_artifact(
|
71
|
-
self.artifacts_wrapper,
|
72
|
-
result,
|
73
|
-
"projects_info.csv",
|
74
|
-
csv_options={"index": False},
|
75
|
-
)
|
76
|
-
|
77
|
-
return (
|
78
|
-
f"You have access to {len(result)} project. "
|
79
|
-
f"Data has been downloaded to the bucket as 'projects_info.csv'."
|
80
|
-
)
|
81
|
-
|
82
|
-
@with_streamlit_logs(tool_name="get_work_items")
|
83
|
-
def get_work_items(
|
84
|
-
self,
|
85
|
-
resolved_after: str,
|
86
|
-
updated_after: str,
|
87
|
-
created_after: str,
|
88
|
-
area: str = "",
|
89
|
-
project_keys: Optional[str] = None,
|
90
|
-
) -> str:
|
91
|
-
"""
|
92
|
-
Get work items from multiple Azure DevOps projects.
|
93
|
-
|
94
|
-
project_keys: str
|
95
|
-
Comma-separated project names.
|
96
|
-
resolved_after: str
|
97
|
-
Date filter for resolved items 'YYYY-MM-DD'.
|
98
|
-
updated_after: str
|
99
|
-
Date filter for updated items 'YYYY-MM-DD'.
|
100
|
-
created_after: str
|
101
|
-
Date filter for created items 'YYYY-MM-DD'.
|
102
|
-
area: str
|
103
|
-
Area path filter (optional).
|
104
|
-
"""
|
105
|
-
project_keys = project_keys or self.project_keys
|
106
|
-
area = area or self.area
|
107
|
-
|
108
|
-
df_work_items = get_work_items_several_projects(
|
109
|
-
project_keys,
|
110
|
-
resolved_after,
|
111
|
-
updated_after,
|
112
|
-
created_after,
|
113
|
-
area=area,
|
114
|
-
ado_search=self.ado_search,
|
115
|
-
)
|
116
|
-
|
117
|
-
save_dataframe_to_artifact(
|
118
|
-
self.artifacts_wrapper,
|
119
|
-
df_work_items,
|
120
|
-
f"{OUTPUT_WORK_ITEMS_FILE}{project_keys}.csv",
|
121
|
-
csv_options={"index_label": "id"},
|
122
|
-
)
|
123
|
-
|
124
|
-
return (
|
125
|
-
f"Work items for {project_keys} have been successfully retrieved "
|
126
|
-
f"and saved to the bucket as '{OUTPUT_WORK_ITEMS}{project_keys}.csv'."
|
127
|
-
)
|
128
|
-
|
129
|
-
async def get_commits(
|
130
|
-
self,
|
131
|
-
since_date: str,
|
132
|
-
project_keys: Optional[str] = None,
|
133
|
-
new_version: bool = True,
|
134
|
-
with_commit_size: bool = True,
|
135
|
-
) -> str:
|
136
|
-
"""
|
137
|
-
Get commits from multiple Azure DevOps projects.
|
138
|
-
|
139
|
-
since_date: str
|
140
|
-
Get commits after this date 'YYYY-MM-DD'.
|
141
|
-
project_keys: str
|
142
|
-
Comma-separated project names.
|
143
|
-
new_version: bool
|
144
|
-
Use new API version.
|
145
|
-
with_commit_size: bool
|
146
|
-
Include commit size info.
|
147
|
-
"""
|
148
|
-
project_keys = project_keys or self.project_keys
|
149
|
-
|
150
|
-
# Await the coroutine to get commits
|
151
|
-
df_commits = await get_commits_several_projects(
|
152
|
-
project_keys,
|
153
|
-
since_date,
|
154
|
-
new_version=new_version,
|
155
|
-
with_commit_size=with_commit_size,
|
156
|
-
ado_search=self.ado_search,
|
157
|
-
)
|
158
|
-
|
159
|
-
save_dataframe_to_artifact(
|
160
|
-
self.artifacts_wrapper,
|
161
|
-
df_commits,
|
162
|
-
f"commits_details_{project_keys}.csv",
|
163
|
-
csv_options={"index_label": "id"},
|
164
|
-
)
|
165
|
-
|
166
|
-
return (
|
167
|
-
f"Commits for {project_keys} have been successfully retrieved "
|
168
|
-
f"and saved to the bucket as 'commits_details_{project_keys}.csv'."
|
169
|
-
)
|
170
|
-
|
171
|
-
def get_merge_requests(
|
172
|
-
self, since_date: str, project_keys: Optional[str] = None
|
173
|
-
) -> str:
|
174
|
-
"""
|
175
|
-
Get pull requests from multiple Azure DevOps projects.
|
176
|
-
|
177
|
-
project_keys: str
|
178
|
-
Comma-separated project names.
|
179
|
-
since_date: str
|
180
|
-
Get PRs after this date 'YYYY-MM-DD'.
|
181
|
-
"""
|
182
|
-
project_keys = project_keys or self.project_keys
|
183
|
-
|
184
|
-
df_prs = get_merge_requests_several_projects(
|
185
|
-
project_keys, since_date, ado_search=self.ado_search
|
186
|
-
)
|
187
|
-
|
188
|
-
save_dataframe_to_artifact(
|
189
|
-
self.artifacts_wrapper,
|
190
|
-
df_prs,
|
191
|
-
f"merge_requests_details_{project_keys}.csv",
|
192
|
-
csv_options={"index": False},
|
193
|
-
)
|
194
|
-
|
195
|
-
return (
|
196
|
-
f"Pull requests for {project_keys} have been successfully retrieved "
|
197
|
-
f"and saved to the bucket as 'merge_requests_details_{project_keys}.csv'."
|
198
|
-
)
|
199
|
-
|
200
|
-
def get_pipelines_runs(
|
201
|
-
self,
|
202
|
-
project_keys: Optional[str] = None,
|
203
|
-
) -> str:
|
204
|
-
"""
|
205
|
-
Get pipeline runs from multiple Azure DevOps projects.
|
206
|
-
|
207
|
-
project_keys: str
|
208
|
-
Comma-separated project names.
|
209
|
-
"""
|
210
|
-
project_keys = project_keys or self.project_keys
|
211
|
-
pipelines_df = get_pipelines_runs_several_projects(project_keys, ado_search=self.ado_search)
|
212
|
-
|
213
|
-
save_dataframe_to_artifact(
|
214
|
-
self.artifacts_wrapper, pipelines_df, f"pipelines_runs_{project_keys}.csv", csv_options={"index": False}
|
215
|
-
)
|
216
|
-
|
217
|
-
return (
|
218
|
-
f"Pipeline runs for {project_keys} have been successfully retrieved "
|
219
|
-
f"and saved to the bucket as 'pipelines_runs_{project_keys}.csv'."
|
220
|
-
)
|
221
|
-
|
222
|
-
def get_available_tools(self) -> list[Dict[str, Any]]:
|
223
|
-
"""Get a list of available tools."""
|
224
|
-
return [
|
225
|
-
{
|
226
|
-
"name": "get_projects_list",
|
227
|
-
"description": self.get_projects_list.__doc__,
|
228
|
-
"ref": self.get_projects_list,
|
229
|
-
"args_schema": NoInputArgs,
|
230
|
-
},
|
231
|
-
{
|
232
|
-
"name": "get_work_items",
|
233
|
-
"description": self.get_work_items.__doc__,
|
234
|
-
"ref": self.get_work_items,
|
235
|
-
"args_schema": GetAdoWorkItemsArgs,
|
236
|
-
},
|
237
|
-
{
|
238
|
-
"name": "get_commits",
|
239
|
-
"description": self.get_commits.__doc__,
|
240
|
-
"ref": self.get_commits,
|
241
|
-
"args_schema": AdoCommitsArgs,
|
242
|
-
},
|
243
|
-
{
|
244
|
-
"name": "get_merge_requests",
|
245
|
-
"description": self.get_merge_requests.__doc__,
|
246
|
-
"ref": self.get_merge_requests,
|
247
|
-
"args_schema": AdoCommitsArgs,
|
248
|
-
},
|
249
|
-
{
|
250
|
-
"name": "get_pipelines_runs",
|
251
|
-
"description": self.get_pipelines_runs.__doc__,
|
252
|
-
"ref": self.get_pipelines_runs,
|
253
|
-
"args_schema": AdoPipelinesArgs,
|
254
|
-
},
|
255
|
-
]
|
256
|
-
|
257
|
-
def run(self, mode: str, *args: Any, **kwargs: Any):
|
258
|
-
for tool in self.get_available_tools():
|
259
|
-
if tool["name"] == mode:
|
260
|
-
return tool["ref"](*args, **kwargs)
|
261
|
-
raise ValueError(f"Unknown mode: {mode}")
|
@@ -1,98 +0,0 @@
|
|
1
|
-
from typing import List, Optional, Literal
|
2
|
-
from pydantic import SecretStr, create_model, BaseModel, ConfigDict, Field
|
3
|
-
|
4
|
-
from langchain_core.tools import BaseTool, BaseToolkit
|
5
|
-
|
6
|
-
from elitea_analyse.github.github_org import GitHubGetOrgLvl
|
7
|
-
from alita_sdk.runtime.clients.client import AlitaClient
|
8
|
-
from alita_sdk.tools.utils import get_max_toolkit_length
|
9
|
-
from alita_sdk.tools.base.tool import BaseAction
|
10
|
-
|
11
|
-
from alita_sdk.runtime.tools.artifact import ArtifactWrapper
|
12
|
-
from .api_wrapper import GitHubAnalyseWrapper
|
13
|
-
|
14
|
-
from ...utils import check_schema
|
15
|
-
|
16
|
-
|
17
|
-
name = "Analyse_Github"
|
18
|
-
|
19
|
-
|
20
|
-
class AnalyseGithub(BaseToolkit):
|
21
|
-
tools: List[BaseTool] = []
|
22
|
-
toolkit_max_length: int = 0
|
23
|
-
|
24
|
-
@staticmethod
|
25
|
-
def toolkit_config_schema() -> type[BaseModel]:
|
26
|
-
selected_tools = {
|
27
|
-
x["name"]: x["args_schema"].schema()
|
28
|
-
for x in GitHubAnalyseWrapper.model_construct().get_available_tools()
|
29
|
-
}
|
30
|
-
AnalyseGithub.toolkit_max_length = get_max_toolkit_length(selected_tools)
|
31
|
-
|
32
|
-
return create_model(
|
33
|
-
"analyse_github",
|
34
|
-
owner=(str, Field(description="GitHub owner name",
|
35
|
-
json_schema_extra={"toolkit_name": True, "max_toolkit_length": AnalyseGithub.toolkit_max_length})),
|
36
|
-
token=(SecretStr, Field(description="Github Access Token", json_schema_extra={"secret": True})),
|
37
|
-
repos=(Optional[str],
|
38
|
-
Field(description="Comma-separated list of GitHub repository names e.g. 'repo1,repo2'", default=None)),
|
39
|
-
artifact_bucket_path=(Optional[str],
|
40
|
-
Field(description="Artifact Bucket Path", default="analyse-github")),
|
41
|
-
selected_tools=(
|
42
|
-
List[Literal[tuple(selected_tools)]], Field(default=[],
|
43
|
-
json_schema_extra={"args_schemas": selected_tools})
|
44
|
-
),
|
45
|
-
__config__=ConfigDict(json_schema_extra={"metadata": {
|
46
|
-
"label": "Analyse_Github",
|
47
|
-
"icon_url": None, # ?? is exists
|
48
|
-
"hidden": True,
|
49
|
-
"sections": {
|
50
|
-
"auth": {
|
51
|
-
"required": True,
|
52
|
-
"subsections": [{"name": "Token", "fields": ["token"]}],
|
53
|
-
}
|
54
|
-
},
|
55
|
-
}
|
56
|
-
})
|
57
|
-
)
|
58
|
-
|
59
|
-
@classmethod
|
60
|
-
def get_toolkit(cls, client: "AlitaClient", selected_tools: list[str], **kwargs):
|
61
|
-
bucket_path = kwargs.get("artifact_bucket_path") or "analyse-github"
|
62
|
-
artifact_wrapper = ArtifactWrapper(client=client, bucket=bucket_path)
|
63
|
-
check_schema(artifact_wrapper)
|
64
|
-
|
65
|
-
owner = kwargs.get("owner")
|
66
|
-
token = kwargs.get("token")
|
67
|
-
|
68
|
-
if not owner or not token:
|
69
|
-
raise ValueError("GitHub owner and token must be provided.")
|
70
|
-
|
71
|
-
git = GitHubGetOrgLvl(owner=owner, token=token)
|
72
|
-
|
73
|
-
github_analyse_wrapper = GitHubAnalyseWrapper(
|
74
|
-
artifacts_wrapper=artifact_wrapper,
|
75
|
-
repos=kwargs.get("repos") or "",
|
76
|
-
git=git
|
77
|
-
)
|
78
|
-
|
79
|
-
selected_tools = selected_tools or []
|
80
|
-
available_tools = github_analyse_wrapper.get_available_tools()
|
81
|
-
|
82
|
-
tools = []
|
83
|
-
for tool in available_tools:
|
84
|
-
if selected_tools and tool["name"] not in selected_tools:
|
85
|
-
continue
|
86
|
-
tools.append(
|
87
|
-
BaseAction(
|
88
|
-
api_wrapper=github_analyse_wrapper,
|
89
|
-
name=tool["name"],
|
90
|
-
description=tool["description"],
|
91
|
-
args_schema=tool["args_schema"],
|
92
|
-
)
|
93
|
-
)
|
94
|
-
|
95
|
-
return cls(tools=tools)
|
96
|
-
|
97
|
-
def get_tools(self):
|
98
|
-
return self.tools
|
@@ -1,166 +0,0 @@
|
|
1
|
-
import logging
|
2
|
-
from typing import Optional, Any
|
3
|
-
from langchain_core.callbacks import dispatch_custom_event
|
4
|
-
from pydantic import BaseModel, Field
|
5
|
-
|
6
|
-
from elitea_analyse.github.github_org import GitHubGetOrgLvl
|
7
|
-
from elitea_analyse.github.main_github import (
|
8
|
-
extract_commits_from_multiple_repos,
|
9
|
-
extract_pull_requests_from_multiple_repos,
|
10
|
-
extract_repositories_list,
|
11
|
-
extract_repositories_extended_data,
|
12
|
-
)
|
13
|
-
|
14
|
-
from alita_sdk.tools.elitea_base import BaseToolApiWrapper
|
15
|
-
from alita_sdk.runtime.utils.save_dataframe import save_dataframe_to_artifact
|
16
|
-
from alita_sdk.runtime.tools.artifact import ArtifactWrapper
|
17
|
-
from alita_sdk.runtime.utils.logging import with_streamlit_logs
|
18
|
-
|
19
|
-
|
20
|
-
logger = logging.getLogger(__name__)
|
21
|
-
|
22
|
-
|
23
|
-
class GetGithubCommitsFromReposArgs(BaseModel):
|
24
|
-
since_after: str = Field( description="Date to filter commits from, in 'YYYY-MM-DD' format." )
|
25
|
-
repos: Optional[str] = Field(
|
26
|
-
description="Comma-separated list of repositories to extract commits from.",
|
27
|
-
default="",
|
28
|
-
)
|
29
|
-
|
30
|
-
|
31
|
-
class GetGithubRepositoriesListArgs(BaseModel):
|
32
|
-
pushed_after: str = Field( description="Date to filter repositories by, in 'YYYY-MM-DD' format." )
|
33
|
-
|
34
|
-
|
35
|
-
class GitHubAnalyseWrapper(BaseToolApiWrapper):
|
36
|
-
artifacts_wrapper: ArtifactWrapper
|
37
|
-
repos: str # Comma-separated list of GitHub repository names e.g. 'repo1,repo2'
|
38
|
-
git: GitHubGetOrgLvl # GitHub client
|
39
|
-
|
40
|
-
class Config:
|
41
|
-
arbitrary_types_allowed = True
|
42
|
-
|
43
|
-
def get_commits_from_repos(self, since_after: str, repos: Optional[str] = None) -> str:
|
44
|
-
"""
|
45
|
-
Extracts commit data from multiple GitHub repositories since the specified date. Saves the result to a CSV file.
|
46
|
-
|
47
|
-
repos : str
|
48
|
-
The string containing repositories names to extract data from, separated by commas.
|
49
|
-
since_date : str
|
50
|
-
The date to start extracting commits from, in 'YYYY-MM-DD' format.
|
51
|
-
"""
|
52
|
-
repos = repos or self.repos
|
53
|
-
df_commits = extract_commits_from_multiple_repos(repos, since_after, git=self.git)
|
54
|
-
|
55
|
-
if df_commits is None or df_commits.empty:
|
56
|
-
return f"No commits found for repositories: {repos} since {since_after}"
|
57
|
-
|
58
|
-
output_filename = f"commits_{repos.replace(',', '_')}.csv"
|
59
|
-
save_dataframe_to_artifact( self.artifacts_wrapper, df_commits, output_filename, {"index": False} )
|
60
|
-
|
61
|
-
return f"GitHub commits data for {repos} saved to {output_filename}"
|
62
|
-
|
63
|
-
def get_pull_requests_from_repos(self, since_after: str, repos: Optional[str] = None) -> str:
|
64
|
-
"""
|
65
|
-
Extracts pull request data from multiple GitHub repositories since the specified date.
|
66
|
-
Saves the result to a CSV file.
|
67
|
-
|
68
|
-
repos: str
|
69
|
-
The string containing repositories names to extract data from, separated by commas.
|
70
|
-
since_date: str
|
71
|
-
The date to start extracting pull requests from, in 'YYYY-MM-DD' format.
|
72
|
-
"""
|
73
|
-
repos = repos or self.repos
|
74
|
-
df_pull_requests = extract_pull_requests_from_multiple_repos(repos, since_after, git=self.git)
|
75
|
-
|
76
|
-
output_filename = f"pull_requests_details_{repos.replace(',', '_')}.csv"
|
77
|
-
save_dataframe_to_artifact( self.artifacts_wrapper, df_pull_requests, output_filename, {"index": False} )
|
78
|
-
|
79
|
-
return f"GitHub pull requests data saved to {output_filename}"
|
80
|
-
|
81
|
-
def get_repositories_list(self, pushed_after: str) -> str:
|
82
|
-
"""
|
83
|
-
Extracts a list of GitHub repositories that were pushed after the specified date.
|
84
|
-
Saves the result to a CSV file.
|
85
|
-
|
86
|
-
pushed_after : str
|
87
|
-
The date to filter repositories by, in 'YYYY-MM-DD' format.
|
88
|
-
"""
|
89
|
-
df_repos = extract_repositories_list(pushed_after, git=self.git)
|
90
|
-
|
91
|
-
output_filename = "github_repos_list.csv"
|
92
|
-
save_dataframe_to_artifact( self.artifacts_wrapper, df_repos, output_filename, {"index": False} )
|
93
|
-
dispatch_custom_event(
|
94
|
-
"thinking_step",
|
95
|
-
data={
|
96
|
-
"message": f"Extracted {len(df_repos)} repositories pushed after {pushed_after}.",
|
97
|
-
"tool_name": "github_repositories_list_extraction",
|
98
|
-
"toolkit": "analyse_github",
|
99
|
-
},
|
100
|
-
)
|
101
|
-
|
102
|
-
return f"GitHub repositories list saved to {output_filename}"
|
103
|
-
|
104
|
-
@with_streamlit_logs(tool_name="get_github_repositories_extended_data")
|
105
|
-
def get_repositories_extended_data(self, pushed_after: str) -> str:
|
106
|
-
"""
|
107
|
-
Extracts extended information about GitHub repositories that were pushed after the specified date.
|
108
|
-
Saves the result to a CSV file.
|
109
|
-
|
110
|
-
pushed_after : str
|
111
|
-
The date to filter repositories by, in 'YYYY-MM-DD' format.
|
112
|
-
"""
|
113
|
-
df_repos_extended = extract_repositories_extended_data(pushed_after, git=self.git)
|
114
|
-
|
115
|
-
output_filename = "github_repos_extended_info.csv"
|
116
|
-
save_dataframe_to_artifact( self.artifacts_wrapper, df_repos_extended, output_filename, {"index": False} )
|
117
|
-
|
118
|
-
dispatch_custom_event(
|
119
|
-
"thinking_step",
|
120
|
-
data={
|
121
|
-
"message": (
|
122
|
-
f"Extracted extended data for {len(df_repos_extended)} repositories "
|
123
|
-
f"pushed after {pushed_after}."
|
124
|
-
),
|
125
|
-
"tool_name": "github_repositories_extended_data_extraction",
|
126
|
-
"toolkit": "analyse_github",
|
127
|
-
},
|
128
|
-
)
|
129
|
-
|
130
|
-
return f"Extended repository info that you have access saved to {output_filename}"
|
131
|
-
|
132
|
-
def get_available_tools(self):
|
133
|
-
"""Get a list of available tools."""
|
134
|
-
return [
|
135
|
-
{
|
136
|
-
"name": "get_commits_from_repos",
|
137
|
-
"description": self.get_commits_from_repos.__doc__,
|
138
|
-
"args_schema": GetGithubCommitsFromReposArgs,
|
139
|
-
"ref": self.get_commits_from_repos,
|
140
|
-
},
|
141
|
-
{
|
142
|
-
"name": "get_pull_requests_from_repos",
|
143
|
-
"description": self.get_pull_requests_from_repos.__doc__,
|
144
|
-
"args_schema": GetGithubCommitsFromReposArgs,
|
145
|
-
"ref": self.get_pull_requests_from_repos,
|
146
|
-
},
|
147
|
-
{
|
148
|
-
"name": "get_repositories_list",
|
149
|
-
"description": self.get_repositories_list.__doc__,
|
150
|
-
"args_schema": GetGithubRepositoriesListArgs,
|
151
|
-
"ref": self.get_repositories_list,
|
152
|
-
},
|
153
|
-
{
|
154
|
-
"name": "get_repositories_extended_data",
|
155
|
-
"description": self.get_repositories_extended_data.__doc__,
|
156
|
-
"args_schema": GetGithubRepositoriesListArgs,
|
157
|
-
"ref": self.get_repositories_extended_data,
|
158
|
-
},
|
159
|
-
]
|
160
|
-
|
161
|
-
def run(self, mode: str, *args: Any, **kwargs: Any):
|
162
|
-
for tool in self.get_available_tools():
|
163
|
-
if tool["name"] == mode:
|
164
|
-
return tool["ref"](*args, **kwargs)
|
165
|
-
|
166
|
-
raise ValueError(f"Unknown mode: {mode}")
|
@@ -1,110 +0,0 @@
|
|
1
|
-
from typing import List, Optional, Literal
|
2
|
-
from elitea_analyse.git.git_search import GitLabV4Search
|
3
|
-
from pydantic import SecretStr, create_model, BaseModel, ConfigDict, Field
|
4
|
-
|
5
|
-
from langchain_core.tools import BaseTool, BaseToolkit
|
6
|
-
|
7
|
-
from alita_sdk.tools.utils import get_max_toolkit_length
|
8
|
-
from alita_sdk.tools.base.tool import BaseAction
|
9
|
-
from alita_sdk.runtime.clients.client import AlitaClient
|
10
|
-
from alita_sdk.runtime.tools.artifact import ArtifactWrapper
|
11
|
-
from .api_wrapper import GitLabAnalyseWrapper
|
12
|
-
|
13
|
-
from ...utils import check_schema
|
14
|
-
|
15
|
-
|
16
|
-
name = "Analyse_GitLab"
|
17
|
-
|
18
|
-
|
19
|
-
class AnalyseGitLab(BaseToolkit):
|
20
|
-
tools: List[BaseTool] = []
|
21
|
-
toolkit_max_length: int = 0
|
22
|
-
|
23
|
-
@staticmethod
|
24
|
-
def toolkit_config_schema() -> type[BaseModel]:
|
25
|
-
selected_tools = {
|
26
|
-
x["name"]: x["args_schema"].schema()
|
27
|
-
for x in GitLabAnalyseWrapper.model_construct().get_available_tools()
|
28
|
-
}
|
29
|
-
AnalyseGitLab.toolkit_max_length = get_max_toolkit_length(selected_tools)
|
30
|
-
|
31
|
-
return create_model(
|
32
|
-
"analyse_gitlab",
|
33
|
-
url=(
|
34
|
-
str,
|
35
|
-
Field(
|
36
|
-
description="GitLab URL (e.g., git.epam.com)",
|
37
|
-
json_schema_extra={"toolkit_name": True, "max_toolkit_length": AnalyseGitLab.toolkit_max_length}
|
38
|
-
)
|
39
|
-
),
|
40
|
-
project_ids=(Optional[str], Field(description="GitLab project ids separated by comma", default=None)),
|
41
|
-
jira_project_keys=(Optional[str],
|
42
|
-
Field(description="GitLab project Jira keys separated by comma", default=None)),
|
43
|
-
token=(SecretStr, Field(description="GitLab Personal Access Token", json_schema_extra={"secret": True})),
|
44
|
-
default_branch_name=(Optional[str], Field(description="Default branch name", default="master")),
|
45
|
-
artifact_bucket_path=(Optional[str], Field(description="Artifact Bucket Path", default="analyse-gitlab")),
|
46
|
-
selected_tools=(
|
47
|
-
List[Literal[tuple(selected_tools)]],
|
48
|
-
Field(default=[], json_schema_extra={"args_schemas": selected_tools})
|
49
|
-
),
|
50
|
-
__config__=ConfigDict(json_schema_extra={"metadata": {
|
51
|
-
"label": "Analyse_GitLab",
|
52
|
-
"icon_url": "gitlab-icon.svg", # if exists
|
53
|
-
"hidden": True,
|
54
|
-
"sections": {
|
55
|
-
"auth": {
|
56
|
-
"required": True,
|
57
|
-
"subsections": [{"name": "Token", "fields": ["token"]}],
|
58
|
-
}
|
59
|
-
},
|
60
|
-
}})
|
61
|
-
)
|
62
|
-
|
63
|
-
@classmethod
|
64
|
-
def get_toolkit(cls, client: "AlitaClient", selected_tools: list[str], **kwargs):
|
65
|
-
bucket_path = kwargs.get("artifact_bucket_path") or "analyse-gitlab"
|
66
|
-
artifact_wrapper = ArtifactWrapper(client=client, bucket=bucket_path)
|
67
|
-
check_schema(artifact_wrapper)
|
68
|
-
|
69
|
-
jira_project_keys = kwargs.get("jira_project_keys") or ""
|
70
|
-
project_ids = kwargs.get("project_ids") or ""
|
71
|
-
url = kwargs.get("url")
|
72
|
-
token = kwargs.get("token")
|
73
|
-
|
74
|
-
if not url or not token:
|
75
|
-
raise ValueError("GitLab URL and token are required.")
|
76
|
-
|
77
|
-
gitlab_search = GitLabV4Search(
|
78
|
-
url=url,
|
79
|
-
default_branch_name=kwargs.get("default_branch_name", "master"),
|
80
|
-
token=token,
|
81
|
-
)
|
82
|
-
|
83
|
-
gitlab_analyse_wrapper = GitLabAnalyseWrapper(
|
84
|
-
artifacts_wrapper=artifact_wrapper,
|
85
|
-
project_ids=project_ids,
|
86
|
-
jira_project_keys=jira_project_keys,
|
87
|
-
gitlab_search=gitlab_search,
|
88
|
-
)
|
89
|
-
|
90
|
-
selected_tools = selected_tools or []
|
91
|
-
available_tools = gitlab_analyse_wrapper.get_available_tools()
|
92
|
-
|
93
|
-
tools = []
|
94
|
-
for tool in available_tools:
|
95
|
-
if selected_tools:
|
96
|
-
if tool["name"] not in selected_tools:
|
97
|
-
continue
|
98
|
-
tools.append(
|
99
|
-
BaseAction(
|
100
|
-
api_wrapper=gitlab_analyse_wrapper,
|
101
|
-
name=tool["name"],
|
102
|
-
description=tool["description"],
|
103
|
-
args_schema=tool["args_schema"],
|
104
|
-
)
|
105
|
-
)
|
106
|
-
|
107
|
-
return cls(tools=tools)
|
108
|
-
|
109
|
-
def get_tools(self):
|
110
|
-
return self.tools
|