Files
crab/pull_requests.py
2025-04-01 15:45:23 +02:00

517 lines
18 KiB
Python
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

from collections import defaultdict
import argparse, os, subprocess, docker
from typing import Any, Callable
from github.Commit import Commit
from github.ContentFile import ContentFile
from github.PullRequest import PullRequest
from github.Repository import Repository
import pandas as pd
from github import Github, GithubException
from pandas.io.common import tarfile
from tqdm import tqdm
from datetime import datetime, timedelta
from dataset import (
Comment,
Dataset,
DatasetEntry,
FileData,
Metadata,
)
from errors import (
CantCheckoutCommitError,
CantEnsureFullHistoryError,
CantFetchPRError,
MultipleFilesError,
NoDiffsAfterError,
NoDiffsBeforeError,
SetupException,
)
from handlers import HandlerException, get_build_handler
from utils import has_only_1_comment, move_github_logging_to_file, clone, run_git_cmd
def get_good_projects(csv_file: str) -> pd.DataFrame:
"""
Extracts the good (the ones that compile and test successfully, and that
have at least one test) from the given file.
Parameters:
csv_file (str): The csv file containing the projects.
Returns:
pd.DataFrame: The good projects.
"""
print(f"Reading {csv_file}...", end="")
df = pd.read_csv(csv_file)
print("Done")
return df.loc[(df['good_repo_for_crab'] == True) & (df['n_tests'] > 0)]
def is_pull_good(pull: PullRequest, verbose: bool = False) -> bool:
comments = pull.get_review_comments()
if pull.user.type == "Bot" or comments.totalCount > 2 or comments.totalCount == 0:
return False
if comments.totalCount == 2:
comment_1, comment_2 = sorted(comments, key=lambda c: c.created_at)
if comment_2.user is not None:
return False # we can't if the second is from the author, so we discard the PR
if comment_1.user is not None and not (
comment_2.in_reply_to_id == comment_1.id # is reply
and comment_2.user.id == pull.user.id # from the author of the PR
):
return False
return has_only_1_comment(pull.get_commits(), pull.get_review_comments(), verbose=verbose)
def ensure_full_history(repo_path: str) -> None:
result = run_git_cmd(["rev-parse", "--is-shallow-repository"], repo_path)
if result.stdout.strip() == "true":
run_git_cmd(["fetch", "--unshallow"], repo_path)
def reset_repo_to_latest_commit(repo_path: str) -> None:
current_branch = run_git_cmd(["rev-parse", "--abbrev-ref", "HEAD"], repo_path).stdout.strip()
run_git_cmd(["reset", "--hard", current_branch], repo_path)
def get_last_commit_before_comments(pr: PullRequest) -> Commit:
comments = list(pr.get_review_comments())
commits = list(pr.get_commits())
comments.sort(key=lambda comment: comment.created_at)
commits.sort(key=lambda commit: commit.commit.author.date)
# remove from the commtis the ones that happened after the first comment
first_comment = comments[0]
for commit in commits[:]:
if commit.commit.author.date > first_comment.created_at:
commits.remove(commit)
return commits[-1]
def get_diffs_before(repo: Repository, pr: PullRequest) -> dict[str, str]:
last_commit_before_comments = get_last_commit_before_comments(pr)
try:
return {
file.filename: file.patch
for file in repo.compare(pr.base.sha, last_commit_before_comments.sha).files
}
except GithubException as e:
raise NoDiffsBeforeError(e)
def get_diffs_after(repo: Repository, pr: PullRequest) -> dict[str, str]:
last_commit_before_comments = get_last_commit_before_comments(pr)
try:
return {
file.filename: file.patch
for file in repo.compare(last_commit_before_comments.sha, pr.merge_commit_sha).files
}
except GithubException as e:
raise NoDiffsAfterError(e)
def checkout(repo_path: str, sha: str, pr_number: int) -> None:
try:
ensure_full_history(repo_path)
except subprocess.CalledProcessError as e:
raise CantEnsureFullHistoryError(e.stderr)
try:
run_git_cmd(["checkout", sha], repo_path)
except subprocess.CalledProcessError:
try:
run_git_cmd(["fetch", "origin", f"pull/{pr_number}/merge"], repo_path)
except subprocess.CalledProcessError as e:
raise CantFetchPRError(e.stderr)
try:
run_git_cmd(["checkout", sha], repo_path)
except subprocess.CalledProcessError as e:
raise CantCheckoutCommitError(e.stderr)
def try_read_file(fname: str) -> str:
if not os.path.exists(fname):
return "" # file was removed after the PR
try:
with open(fname, "r", encoding="utf-8") as f:
return f.read()
except UnicodeDecodeError:
return "Binary file (from filesystem), to be ignored"
except IsADirectoryError:
return "File listed in PR is a directory (likely a submodule), to be ignored"
def get_files(pr: PullRequest, repo: Repository, repo_path: str) -> dict[str, FileData]:
ret = {}
for file in pr.get_files():
try:
contents = repo.get_contents(file.filename, ref=pr.base.sha)
assert isinstance(
contents, ContentFile
), f"Multiple files with the same name {file.filename} in base sha {pr.base.sha} ({contents})"
contents_before = contents.decoded_content.decode()
except AssertionError as e:
raise MultipleFilesError(e)
except UnicodeError as e:
contents_before = "Binary content (from API), to be ignored"
except Exception as e:
contents_before = "" # file didn't exist before the PR
try:
contents = repo.get_contents(file.filename, ref=pr.merge_commit_sha)
assert isinstance(
contents, ContentFile
), f"Multiple files with the same name {file.filename} in merge commit sha {pr.base.sha} ({contents})"
contents_after = contents.decoded_content.decode()
except AssertionError as e:
raise MultipleFilesError(e)
except UnicodeError as e:
contents_after = "Binary content (from API), to be ignored"
except Exception as e:
checkout(repo_path, pr.merge_commit_sha, pr.number)
contents_after = try_read_file(os.path.join(repo_path, file.filename))
ret[file.filename] = FileData(
is_code_related=file.filename.endswith('.java'),
coverage={},
content_before_pr=contents_before,
content_after_pr=contents_after,
)
return ret
def get_comments(pr: PullRequest) -> list[Comment]:
ret = []
for comment in pr.get_review_comments():
comment_to_add = Comment(
body=comment.body,
file=comment.path,
from_=comment.start_line if comment.start_line else comment.line,
to=comment.line,
)
if comment_to_add.from_ is None or comment_to_add.to is None:
comment_to_add.to = comment.original_line
comment_to_add.from_ = comment.original_start_line
ret.append(comment_to_add)
return ret
def archive_repo(
repo_path: str, repo_name: str, pr_number: int, destination: str, post_fix: str
) -> None:
"""
Archives the repo at the specified path, including only the files tracked by git.
The archive is stored in the destination directory with a filename based on the PR number.
"""
if not os.path.exists(destination):
os.makedirs(destination)
archive_name = f"{repo_name.replace('/', '_')}_{pr_number}_{post_fix}.tar.gz"
archive_path = os.path.join(destination, archive_name)
result = run_git_cmd(["ls-files"], repo_path)
tracked_files = result.stdout.strip().split("\n")
with tarfile.open(archive_path, "w:gz") as tar:
for file in tracked_files:
full_path = os.path.join(repo_path, file)
if os.path.exists(full_path):
tar.add(full_path, arcname=file)
def process_pull(
repo: Repository,
pr: PullRequest,
dataset: Dataset,
repos_dir: str,
archive_destination: str,
cache: dict[str, dict[int, DatasetEntry]] = {},
):
if pr.number in cache.get(repo.full_name, set()):
dataset.entries.append(cache[repo.full_name][pr.number])
return
entry = DatasetEntry(
metadata=Metadata(
repo.full_name,
pr.number,
pr.title,
pr.body,
pr.merge_commit_sha,
reason_for_failure="Was still being processed",
),
files={},
diffs_before={},
comments=[],
diffs_after={},
)
dataset.entries.append(entry)
comments = pr.get_review_comments()
assert comments.totalCount == 1
comment = comments[0]
commented_file_path = comment.path
repo_path = os.path.join(repos_dir, repo.full_name)
build_handler = None
setup_steps = [
(
"Getting diffs before the first commit...",
lambda: entry.diffs_before.update(get_diffs_before(repo, pr)),
),
(
"Getting diffs after the first commit...",
lambda: entry.diffs_after.update(get_diffs_after(repo, pr)),
),
("Cloning the repo...", lambda: clone(repo.full_name, repos_dir)),
(
"Getting the files...",
lambda: entry.files.update(get_files(pr, repo, repo_path)),
),
(
"Getting the comments...",
lambda: entry.comments.extend(get_comments(pr)),
),
("Checkout out base commit...", lambda: checkout(repo_path, pr.base.sha, pr.number)),
(
"Archiving the repo...",
lambda: archive_repo(repo_path, repo.full_name, pr.number, archive_destination, "base"),
),
(
"Checkout out merge commit...",
lambda: checkout(repo_path, pr.merge_commit_sha, pr.number),
),
(
"Archiving the repo...",
lambda: archive_repo(
repo_path, repo.full_name, pr.number, archive_destination, "merged"
),
),
]
pbar = tqdm(total=len(setup_steps) + 6, desc="Processing PR", leave=False)
for message, action in setup_steps:
pbar.set_postfix(
{
"doing": message,
"started at": datetime.now().strftime("%d/%m, %H:%M:%S"),
}
)
try:
action()
except SetupException as e:
entry.metadata.last_cmd_error_msg = str(e)
entry.metadata.reason_for_failure = e.reason_for_failure
entry.metadata.successful = False
return
pbar.update(1)
try:
pbar.set_postfix(
{
"doing": "Setting up build handler...",
"started at": datetime.now().strftime("%d/%m, %H:%M:%S"),
}
)
build_handler = get_build_handler(repos_dir, repo.full_name)
pbar.update(1)
entry.metadata.build_system = build_handler.get_type()
build_handler.set_client(docker_client)
except SetupException as e:
entry.metadata.last_cmd_error_msg = str(e)
entry.metadata.reason_for_failure = e.reason_for_failure
entry.metadata.successful = False
return
def _check_coverages():
for coverage_file, coverage in build_handler.check_coverage(commented_file_path):
entry.files[commented_file_path].coverage[coverage_file] = coverage
steps = [
("Checking for tests...", build_handler.check_for_tests),
("Compiling...", build_handler.compile_repo),
("Running tests...", build_handler.test_repo),
("Generating coverage...", build_handler.generate_coverage_report),
("Checking coverage...", _check_coverages),
]
if all(not comment.file.endswith(".java") for comment in entry.comments):
# if the commented files are all not code related, why bother compiling and testing the code?
pbar.update(5)
if entry.metadata.successful:
entry.metadata.reason_for_failure = "Valid PR! But isn't code related though."
return
with build_handler:
try:
for message, action in steps:
pbar.set_postfix(
{
"doing": message,
"started at": datetime.now().strftime("%d/%m, %H:%M:%S"),
}
)
action()
pbar.update(1)
except HandlerException as e:
entry.metadata.last_cmd_error_msg = str(e)
entry.metadata.reason_for_failure = e.reason_for_failure
entry.metadata.successful = False
finally:
build_handler.clean_repo()
reset_repo_to_latest_commit(repo_path)
if entry.metadata.successful:
entry.metadata.reason_for_failure = "Valid PR!" # was set to 'still processing', since it's done being processed and was successful, there are no reasons for failure
def process_repo(
repo_name: str,
dataset: Dataset,
repos_dir: str,
archive_destination: str,
cache: dict[str, dict[int, DatasetEntry]] = {},
):
repo = g.get_repo(repo_name)
already_seen_prs = set()
if repo.full_name in cache:
dataset.entries.extend(cache[repo.full_name].values())
already_seen_prs = set(cache[repo.full_name].keys())
dataset.to_json(args.output)
prs = repo.get_pulls(state="closed")
n_good_prs = 0
with tqdm(total=prs.totalCount, desc="Processing prs", leave=False) as pbar:
for pr in prs:
pbar.set_postfix({"pr": pr.number, "# new good found": n_good_prs})
try:
if pr.merged_at is None or pr.number in already_seen_prs or not is_pull_good(pr):
continue
n_good_prs += 1
process_pull(repo, pr, dataset, repos_dir, archive_destination, cache)
dataset.to_json(args.output)
except Exception as e:
tqdm.write(f"[ERROR] PR #{pr.number} in {repo.full_name}. {type(e)}: {e}")
finally:
pbar.update(1)
def process_repos(
df: pd.DataFrame,
dataset: Dataset,
repos_dir: str,
archive_destination: str,
cache: dict[str, dict[int, DatasetEntry]] = {},
):
"""
Processes the repos in the given csv file, extracting the good ones and
creating the "triplets" for the dataset.
Parameters:
csv_file (str): The csv file containing the projects.
dataset (Dataset): The dataset in which the triplets will be stored.
Passing it by reference in order have the latest information, in case of an error
verbose (bool): Whether to be verbose or not
"""
with tqdm(total=len(df), desc="Processing repos") as pbar:
for _, row in df.iterrows():
repo_name = row["name"]
assert isinstance(repo_name, str)
pbar.set_postfix(
{
"repo": repo_name,
"started at": datetime.now().strftime("%d/%m, %H:%M:%S"),
"# triplets": f"{len(dataset)}/{len(dataset.entries)} ({len(dataset)/len(dataset.entries) if len(dataset.entries) > 0 else 0:.2%})",
}
)
process_repo(repo_name, dataset, repos_dir, archive_destination, cache)
pbar.update(1)
if __name__ == "__main__":
parser = argparse.ArgumentParser(description='Creates the triplets for the CRAB dataset.')
parser.add_argument(
'csv_file',
type=str,
help='The csv file containing the projects (the results from clone_repos.py).',
)
parser.add_argument(
'-o',
'--output',
type=str,
default="./dataset.json",
help='The file in which the dataset will be contained. Default is "./dataset.json"',
)
parser.add_argument(
'-r',
'--repos',
type=str,
default="./results/",
help='The directory in which the repos were cloned (will be cloned if they aren\'t there already). Default: "./results/"',
)
parser.add_argument(
'-c',
'--cache',
type=str,
help="The name of the output file from another run of this script. This is for when the script unexpectedly got interrupted and you want to resume from where you left off.",
)
parser.add_argument(
"-a",
"--archive-destination",
type=str,
default="./dataset/archives",
help="The directory in which the repos will be archived. Default is './dataset/archives'.",
)
# parser.add_argument('-v', '--verbose', action='store_true', help='Prints the number of good projects.')
parser.add_argument(
"--only-repo",
type=str,
metavar="OWNER/NAME",
help="Run the script on a single repo (format: 'owner/name'). If not set, all repos in '--repos' CSV are processed.",
)
parser.add_argument(
"--cache-requests",
action="store_true",
help="Caches GitHub API requests in a SQLite file using 'requests_cache' (see optional-requirements.txt). Useful for faster reruns if the script crashes or youre tweaking it. Might produce stale data.",
)
args = parser.parse_args()
if args.cache_requests:
import requests_cache
requests_cache.install_cache('github_cache', expire_after=timedelta(weeks=2))
g = Github(os.environ["GITHUB_AUTH_TOKEN_CRAB"], seconds_between_requests=0)
docker_client = docker.from_env()
move_github_logging_to_file()
df = get_good_projects(args.csv_file)
if args.only_repo is not None:
df = df.loc[df["name"] == args.only_repo]
cache: dict[str, dict[int, DatasetEntry]] = defaultdict(dict)
if args.cache is not None:
cache_dataset = Dataset.from_json(args.cache)
for cache_entry in cache_dataset.entries:
cache[cache_entry.metadata.repo][cache_entry.metadata.pr_number] = cache_entry
dataset = Dataset()
try:
process_repos(df, dataset, args.repos, args.archive_destination, cache)
finally:
dataset.to_json(args.output)