|
|
|
|
@ -1,4 +1,5 @@
|
|
|
|
|
import json
|
|
|
|
|
import urllib.parse
|
|
|
|
|
from datetime import datetime, timedelta
|
|
|
|
|
from typing import Any, Union
|
|
|
|
|
|
|
|
|
|
@ -13,13 +14,14 @@ class GitlabCommitsTool(BuiltinTool):
|
|
|
|
|
self, user_id: str, tool_parameters: dict[str, Any]
|
|
|
|
|
) -> Union[ToolInvokeMessage, list[ToolInvokeMessage]]:
|
|
|
|
|
project = tool_parameters.get("project", "")
|
|
|
|
|
repository = tool_parameters.get("repository", "")
|
|
|
|
|
employee = tool_parameters.get("employee", "")
|
|
|
|
|
start_time = tool_parameters.get("start_time", "")
|
|
|
|
|
end_time = tool_parameters.get("end_time", "")
|
|
|
|
|
change_type = tool_parameters.get("change_type", "all")
|
|
|
|
|
|
|
|
|
|
if not project:
|
|
|
|
|
return self.create_text_message("Project is required")
|
|
|
|
|
if not project and not repository:
|
|
|
|
|
return self.create_text_message("Either project or repository is required")
|
|
|
|
|
|
|
|
|
|
if not start_time:
|
|
|
|
|
start_time = (datetime.utcnow() - timedelta(days=1)).isoformat()
|
|
|
|
|
@ -35,91 +37,105 @@ class GitlabCommitsTool(BuiltinTool):
|
|
|
|
|
site_url = "https://gitlab.com"
|
|
|
|
|
|
|
|
|
|
# Get commit content
|
|
|
|
|
result = self.fetch(user_id, site_url, access_token, project, employee, start_time, end_time, change_type)
|
|
|
|
|
if repository:
|
|
|
|
|
result = self.fetch_commits(
|
|
|
|
|
site_url, access_token, repository, employee, start_time, end_time, change_type, is_repository=True
|
|
|
|
|
)
|
|
|
|
|
else:
|
|
|
|
|
result = self.fetch_commits(
|
|
|
|
|
site_url, access_token, project, employee, start_time, end_time, change_type, is_repository=False
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
return [self.create_json_message(item) for item in result]
|
|
|
|
|
|
|
|
|
|
def fetch(
|
|
|
|
|
def fetch_commits(
|
|
|
|
|
self,
|
|
|
|
|
user_id: str,
|
|
|
|
|
site_url: str,
|
|
|
|
|
access_token: str,
|
|
|
|
|
project: str,
|
|
|
|
|
employee: str = None,
|
|
|
|
|
start_time: str = "",
|
|
|
|
|
end_time: str = "",
|
|
|
|
|
change_type: str = "",
|
|
|
|
|
identifier: str,
|
|
|
|
|
employee: str,
|
|
|
|
|
start_time: str,
|
|
|
|
|
end_time: str,
|
|
|
|
|
change_type: str,
|
|
|
|
|
is_repository: bool,
|
|
|
|
|
) -> list[dict[str, Any]]:
|
|
|
|
|
domain = site_url
|
|
|
|
|
headers = {"PRIVATE-TOKEN": access_token}
|
|
|
|
|
results = []
|
|
|
|
|
|
|
|
|
|
try:
|
|
|
|
|
# Get all of projects
|
|
|
|
|
url = f"{domain}/api/v4/projects"
|
|
|
|
|
response = requests.get(url, headers=headers)
|
|
|
|
|
response.raise_for_status()
|
|
|
|
|
projects = response.json()
|
|
|
|
|
|
|
|
|
|
filtered_projects = [p for p in projects if project == "*" or p["name"] == project]
|
|
|
|
|
|
|
|
|
|
for project in filtered_projects:
|
|
|
|
|
project_id = project["id"]
|
|
|
|
|
project_name = project["name"]
|
|
|
|
|
print(f"Project: {project_name}")
|
|
|
|
|
|
|
|
|
|
# Get all of project commits
|
|
|
|
|
commits_url = f"{domain}/api/v4/projects/{project_id}/repository/commits"
|
|
|
|
|
params = {"since": start_time, "until": end_time}
|
|
|
|
|
if employee:
|
|
|
|
|
params["author"] = employee
|
|
|
|
|
|
|
|
|
|
commits_response = requests.get(commits_url, headers=headers, params=params)
|
|
|
|
|
commits_response.raise_for_status()
|
|
|
|
|
commits = commits_response.json()
|
|
|
|
|
|
|
|
|
|
for commit in commits:
|
|
|
|
|
commit_sha = commit["id"]
|
|
|
|
|
author_name = commit["author_name"]
|
|
|
|
|
|
|
|
|
|
if is_repository:
|
|
|
|
|
# URL encode the repository path
|
|
|
|
|
encoded_identifier = urllib.parse.quote(identifier, safe="")
|
|
|
|
|
commits_url = f"{domain}/api/v4/projects/{encoded_identifier}/repository/commits"
|
|
|
|
|
else:
|
|
|
|
|
# Get all projects
|
|
|
|
|
url = f"{domain}/api/v4/projects"
|
|
|
|
|
response = requests.get(url, headers=headers)
|
|
|
|
|
response.raise_for_status()
|
|
|
|
|
projects = response.json()
|
|
|
|
|
|
|
|
|
|
filtered_projects = [p for p in projects if identifier == "*" or p["name"] == identifier]
|
|
|
|
|
|
|
|
|
|
for project in filtered_projects:
|
|
|
|
|
project_id = project["id"]
|
|
|
|
|
project_name = project["name"]
|
|
|
|
|
print(f"Project: {project_name}")
|
|
|
|
|
|
|
|
|
|
commits_url = f"{domain}/api/v4/projects/{project_id}/repository/commits"
|
|
|
|
|
|
|
|
|
|
params = {"since": start_time, "until": end_time}
|
|
|
|
|
if employee:
|
|
|
|
|
params["author"] = employee
|
|
|
|
|
|
|
|
|
|
commits_response = requests.get(commits_url, headers=headers, params=params)
|
|
|
|
|
commits_response.raise_for_status()
|
|
|
|
|
commits = commits_response.json()
|
|
|
|
|
|
|
|
|
|
for commit in commits:
|
|
|
|
|
commit_sha = commit["id"]
|
|
|
|
|
author_name = commit["author_name"]
|
|
|
|
|
|
|
|
|
|
if is_repository:
|
|
|
|
|
diff_url = f"{domain}/api/v4/projects/{encoded_identifier}/repository/commits/{commit_sha}/diff"
|
|
|
|
|
else:
|
|
|
|
|
diff_url = f"{domain}/api/v4/projects/{project_id}/repository/commits/{commit_sha}/diff"
|
|
|
|
|
diff_response = requests.get(diff_url, headers=headers)
|
|
|
|
|
diff_response.raise_for_status()
|
|
|
|
|
diffs = diff_response.json()
|
|
|
|
|
|
|
|
|
|
for diff in diffs:
|
|
|
|
|
# Calculate code lines of changed
|
|
|
|
|
added_lines = diff["diff"].count("\n+")
|
|
|
|
|
removed_lines = diff["diff"].count("\n-")
|
|
|
|
|
total_changes = added_lines + removed_lines
|
|
|
|
|
|
|
|
|
|
if change_type == "new":
|
|
|
|
|
if added_lines > 1:
|
|
|
|
|
final_code = "".join(
|
|
|
|
|
[
|
|
|
|
|
line[1:]
|
|
|
|
|
for line in diff["diff"].split("\n")
|
|
|
|
|
if line.startswith("+") and not line.startswith("+++")
|
|
|
|
|
]
|
|
|
|
|
)
|
|
|
|
|
results.append(
|
|
|
|
|
{"commit_sha": commit_sha, "author_name": author_name, "diff": final_code}
|
|
|
|
|
)
|
|
|
|
|
else:
|
|
|
|
|
if total_changes > 1:
|
|
|
|
|
final_code = "".join(
|
|
|
|
|
[
|
|
|
|
|
line[1:]
|
|
|
|
|
for line in diff["diff"].split("\n")
|
|
|
|
|
if (line.startswith("+") or line.startswith("-"))
|
|
|
|
|
and not line.startswith("+++")
|
|
|
|
|
and not line.startswith("---")
|
|
|
|
|
]
|
|
|
|
|
)
|
|
|
|
|
final_code_escaped = json.dumps(final_code)[1:-1] # Escape the final code
|
|
|
|
|
results.append(
|
|
|
|
|
{"commit_sha": commit_sha, "author_name": author_name, "diff": final_code_escaped}
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
diff_response = requests.get(diff_url, headers=headers)
|
|
|
|
|
diff_response.raise_for_status()
|
|
|
|
|
diffs = diff_response.json()
|
|
|
|
|
|
|
|
|
|
for diff in diffs:
|
|
|
|
|
# Calculate code lines of changes
|
|
|
|
|
added_lines = diff["diff"].count("\n+")
|
|
|
|
|
removed_lines = diff["diff"].count("\n-")
|
|
|
|
|
total_changes = added_lines + removed_lines
|
|
|
|
|
|
|
|
|
|
if change_type == "new":
|
|
|
|
|
if added_lines > 1:
|
|
|
|
|
final_code = "".join(
|
|
|
|
|
[
|
|
|
|
|
line[1:]
|
|
|
|
|
for line in diff["diff"].split("\n")
|
|
|
|
|
if line.startswith("+") and not line.startswith("+++")
|
|
|
|
|
]
|
|
|
|
|
)
|
|
|
|
|
results.append({"commit_sha": commit_sha, "author_name": author_name, "diff": final_code})
|
|
|
|
|
else:
|
|
|
|
|
if total_changes > 1:
|
|
|
|
|
final_code = "".join(
|
|
|
|
|
[
|
|
|
|
|
line[1:]
|
|
|
|
|
for line in diff["diff"].split("\n")
|
|
|
|
|
if (line.startswith("+") or line.startswith("-"))
|
|
|
|
|
and not line.startswith("+++")
|
|
|
|
|
and not line.startswith("---")
|
|
|
|
|
]
|
|
|
|
|
)
|
|
|
|
|
final_code_escaped = json.dumps(final_code)[1:-1] # Escape the final code
|
|
|
|
|
results.append(
|
|
|
|
|
{"commit_sha": commit_sha, "author_name": author_name, "diff": final_code_escaped}
|
|
|
|
|
)
|
|
|
|
|
except requests.RequestException as e:
|
|
|
|
|
print(f"Error fetching data from GitLab: {e}")
|
|
|
|
|
|
|
|
|
|
|