blob: ef2df11048cafc596da4098634899855f92672b4 [file] [log] [blame] [edit]
#!/usr/bin/env python3
#
# Copyright (C) 2013-2015 The CyanogenMod Project
# (C) 2017-2024 The LineageOS Project
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
#
# Run repopick.py -h for a description of this utility.
#
import argparse
import json
import os
import re
import subprocess
import sys
import textwrap
import urllib.parse
import urllib.request
from collections import defaultdict
from concurrent.futures import ThreadPoolExecutor
from functools import cmp_to_key, partial
from xml.etree import ElementTree
# cmp() is not available in Python 3, define it manually
# See https://docs.python.org/3.0/whatsnew/3.0.html#ordering-comparisons
def cmp(a, b):
return (a > b) - (a < b)
# Verifies whether pathA is a subdirectory (or the same) as pathB
def is_subdir(a, b):
a = os.path.realpath(a) + "/"
b = os.path.realpath(b) + "/"
return b == a[: len(b)]
def fetch_query_via_ssh(remote_url, query):
"""Given a remote_url and a query, return the list of changes that fit it
This function is slightly messy - the ssh api does not return data in the same structure as the HTTP REST API
We have to get the data, then transform it to match what we're expecting from the HTTP RESET API
"""
if remote_url.count(":") == 2:
(_, userhost, port) = remote_url.split(":")
userhost = userhost[2:]
elif remote_url.count(":") == 1:
(_, userhost) = remote_url.split(":")
userhost = userhost[2:]
port = "29418"
else:
raise Exception("Malformed URI: Expecting ssh://[user@]host[:port]")
out = subprocess.check_output(
[
"ssh",
"-x",
"-p",
port,
userhost,
"gerrit",
"query",
"--format",
"JSON",
"--patch-sets",
"--current-patch-set",
query,
],
text=True,
)
reviews = []
for line in out.split("\n"):
try:
data = json.loads(line)
# make our data look like the http rest api data
review = {
"branch": data["branch"],
"change_id": data["id"],
"current_revision": data["currentPatchSet"]["revision"],
"number": int(data["number"]),
"revisions": {
patch_set["revision"]: {
"_number": int(patch_set["number"]),
"fetch": {
"ssh": {
"ref": patch_set["ref"],
"url": "ssh://{0}:{1}/{2}".format(
userhost, port, data["project"]
),
}
},
"commit": {
"parents": [
{"commit": parent} for parent in patch_set["parents"]
]
},
}
for patch_set in data["patchSets"]
},
"subject": data["subject"],
"project": data["project"],
"status": data["status"],
}
reviews.append(review)
except Exception:
pass
return reviews
def build_query_url(remote_url, query, auth):
p = urllib.parse.urlparse(remote_url)._asdict()
p["path"] = ("/a" if auth else "") + "/changes"
p["query"] = urllib.parse.urlencode(
{
"q": query,
"o": ["CURRENT_REVISION", "ALL_REVISIONS", "ALL_COMMITS"],
},
doseq=True,
)
return urllib.parse.urlunparse(urllib.parse.ParseResult(**p))
def fetch_query_via_http(remote_url, query, auth=True):
"""Given a query, fetch the change numbers via http"""
if auth:
gerritrc = os.path.expanduser("~/.gerritrc")
username = password = ""
if os.path.isfile(gerritrc):
with open(gerritrc, "r") as f:
for line in f:
parts = line.rstrip().split("|")
if parts[0] in remote_url:
username, password = parts[1], parts[2]
if username and password:
url = build_query_url(remote_url, query, auth)
password_mgr = urllib.request.HTTPPasswordMgrWithDefaultRealm()
password_mgr.add_password(None, url, username, password)
auth_handler = urllib.request.HTTPBasicAuthHandler(password_mgr)
opener = urllib.request.build_opener(auth_handler)
response = opener.open(url)
if response.getcode() != 200:
# They didn't get good authorization or data, Let's try the old way
return fetch_query_via_http(remote_url, query, False)
else:
return fetch_query_via_http(remote_url, query, False)
else:
url = build_query_url(remote_url, query, auth)
response = urllib.request.urlopen(url)
data = response.read().decode("utf-8")
reviews = json.loads(data[5:])
for review in reviews:
review["number"] = review.pop("_number")
return reviews
def fetch_query(remote_url, query):
"""Wrapper for fetch_query_via_proto functions"""
if remote_url[0:3] == "ssh":
return fetch_query_via_ssh(remote_url, query)
elif remote_url[0:4] == "http":
return fetch_query_via_http(remote_url, query)
else:
raise Exception(
"Gerrit URL should be in the form http[s]://hostname/ or ssh://[user@]host[:port]"
)
def is_closed(status):
return status not in ("OPEN", "NEW", "DRAFT")
def is_omnirom_gerrit(remote_url):
p = urllib.parse.urlparse(remote_url)
return p.hostname == "gerrit.omnirom.org"
def commit_exists(project_path, revision):
return (
subprocess.call(
["git", "cat-file", "-e", revision],
cwd=project_path,
stderr=subprocess.DEVNULL,
)
== 0
)
def main():
parser = argparse.ArgumentParser(
formatter_class=argparse.RawDescriptionHelpFormatter,
description=textwrap.dedent(
"""\
repopick.py is a utility to simplify the process of cherry picking
patches from OmniRom's Gerrit instance (or any gerrit instance of your choosing)
Given a list of change numbers, repopick will cd into the project path
and cherry pick the latest patch available.
With the --start-branch argument, the user can specify that a branch
should be created before cherry picking. This is useful for
cherry-picking many patches into a common branch which can be easily
abandoned later (good for testing other's changes.)
The --abandon-first argument, when used in conjunction with the
--start-branch option, will cause repopick to abandon the specified
branch in all repos first before performing any cherry picks."""
),
)
parser.add_argument(
"change_number",
nargs="*",
help="change number to cherry pick. Use {change number}/{patchset number} to get a specific revision.",
)
parser.add_argument(
"-i",
"--ignore-missing",
action="store_true",
help="do not error out if a patch applies to a missing directory",
)
parser.add_argument(
"-s",
"--start-branch",
nargs=1,
metavar="",
help="start the specified branch before cherry picking",
)
parser.add_argument(
"-r",
"--reset",
action="store_true",
help="reset to initial state (abort cherry-pick) if there is a conflict",
)
parser.add_argument(
"-a",
"--abandon-first",
action="store_true",
help="before cherry picking, abandon the branch specified in --start-branch",
)
parser.add_argument(
"-b",
"--auto-branch",
action="store_true",
help='shortcut to "--start-branch auto --abandon-first --ignore-missing"',
)
parser.add_argument(
"-q", "--quiet", action="store_true", help="print as little as possible"
)
parser.add_argument(
"-v",
"--verbose",
action="store_true",
help="print extra information to aid in debug",
)
parser.add_argument(
"-f",
"--force",
action="store_true",
help="force cherry pick even if change is closed",
)
parser.add_argument(
"-p", "--pull", action="store_true", help="execute pull instead of cherry-pick"
)
parser.add_argument(
"-P", "--path", metavar="", help="use the specified path for the change"
)
parser.add_argument(
"-t", "--topic", metavar="", help="pick all commits from a specified topic"
)
parser.add_argument(
"-Q", "--query", metavar="", help="pick all commits using the specified query"
)
parser.add_argument(
"-g",
"--gerrit",
default="https://gerrit.omnirom.org",
metavar="",
help="Gerrit Instance to use. Form proto://[user@]host[:port]",
)
parser.add_argument(
"-e",
"--exclude",
nargs=1,
metavar="",
help="exclude a list of commit numbers separated by a ,",
)
parser.add_argument(
"-c",
"--check-picked",
type=int,
default=10,
metavar="",
help="pass the amount of commits to check for already picked changes",
)
parser.add_argument(
"-j",
"--jobs",
type=int,
default=4,
metavar="",
help="max number of changes to pick in parallel",
)
args = parser.parse_args()
if not args.start_branch and args.abandon_first:
parser.error(
"if --abandon-first is set, you must also give the branch name with --start-branch"
)
if args.auto_branch:
args.abandon_first = True
args.ignore_missing = True
if not args.start_branch:
args.start_branch = ["auto"]
if args.quiet and args.verbose:
parser.error("--quiet and --verbose cannot be specified together")
if (1 << bool(args.change_number) << bool(args.topic) << bool(args.query)) != 2:
parser.error(
"One (and only one) of change_number, topic, and query are allowed"
)
# Change current directory to the top of the tree
if "ANDROID_BUILD_TOP" in os.environ:
top = os.environ["ANDROID_BUILD_TOP"]
if not is_subdir(os.getcwd(), top):
sys.stderr.write(
"ERROR: You must run this tool from within $ANDROID_BUILD_TOP!\n"
)
sys.exit(1)
os.chdir(os.environ["ANDROID_BUILD_TOP"])
# Sanity check that we are being run from the top level of the tree
if not os.path.isdir(".repo"):
sys.stderr.write(
"ERROR: No .repo directory found. Please run this from the top of your tree.\n"
)
sys.exit(1)
# If --abandon-first is given, abandon the branch before starting
if args.abandon_first:
# Determine if the branch already exists; skip the abandon if it does not
plist = subprocess.check_output(["repo", "info"], text=True)
needs_abandon = False
for pline in plist.splitlines():
matchObj = re.match(r"Local Branches.*\[(.*)\]", pline)
if matchObj:
local_branches = re.split(r"\s*,\s*", matchObj.group(1))
if any(args.start_branch[0] in s for s in local_branches):
needs_abandon = True
if needs_abandon:
# Perform the abandon only if the branch already exists
if not args.quiet:
print("Abandoning branch: %s" % args.start_branch[0])
subprocess.run(["repo", "abandon", args.start_branch[0]])
if not args.quiet:
print("")
# Get the master manifest from repo
# - convert project name and revision to a path
project_name_to_data = {}
manifest = subprocess.check_output(["repo", "manifest"], text=True)
xml_root = ElementTree.fromstring(manifest)
projects = xml_root.findall("project")
remotes = xml_root.findall("remote")
default_revision = xml_root.findall("default")[0].get("revision")
if not default_revision:
raise ValueError("Failed to get revision from manifest")
# dump project data into the a list of dicts with the following data:
# {project: {path, revision}}
for project in projects:
name = project.get("name")
# when name and path are equal, "repo manifest" doesn't return a path at all, so fall back to name
path = project.get("path", name)
revision = project.get("upstream")
if revision is None:
for remote in remotes:
if remote.get("name") == project.get("remote"):
revision = remote.get("revision")
if revision is None:
revision = project.get("revision", default_revision)
if name not in project_name_to_data:
project_name_to_data[name] = {}
revision = revision.split("refs/heads/")[-1]
project_name_to_data[name][revision] = path
def cmp_reviews(review_a, review_b):
current_a = review_a["current_revision"]
parents_a = [
r["commit"] for r in review_a["revisions"][current_a]["commit"]["parents"]
]
current_b = review_b["current_revision"]
parents_b = [
r["commit"] for r in review_b["revisions"][current_b]["commit"]["parents"]
]
if current_a in parents_b:
return -1
elif current_b in parents_a:
return 1
else:
return cmp(review_a["number"], review_b["number"])
# get data on requested changes
if args.topic:
reviews = fetch_query(args.gerrit, "topic:{0}".format(args.topic))
change_numbers = [
str(r["number"]) for r in sorted(reviews, key=cmp_to_key(cmp_reviews))
]
elif args.query:
reviews = fetch_query(args.gerrit, args.query)
change_numbers = [
str(r["number"]) for r in sorted(reviews, key=cmp_to_key(cmp_reviews))
]
else:
change_url_re = re.compile(r"https?://.+?/([0-9]+(?:/[0-9]+)?)/?")
change_numbers = []
for c in args.change_number:
change_number = change_url_re.findall(c)
if change_number:
change_numbers.extend(change_number)
elif "-" in c:
templist = c.split("-")
for i in range(int(templist[0]), int(templist[1]) + 1):
change_numbers.append(str(i))
else:
change_numbers.append(c)
reviews = fetch_query(
args.gerrit,
" OR ".join("change:{0}".format(x.split("/")[0]) for x in change_numbers),
)
# make list of things to actually merge
mergables = defaultdict(list)
# If --exclude is given, create the list of commits to ignore
exclude = []
if args.exclude:
exclude = args.exclude[0].split(",")
for change in change_numbers:
patchset = None
if "/" in change:
(change, patchset) = change.split("/")
if change in exclude:
continue
change = int(change)
if patchset is not None:
patchset = int(patchset)
review = next((x for x in reviews if x["number"] == change), None)
if review is None:
print("Change %d not found, skipping" % change)
continue
# Check if change is open and exit if it's not, unless -f is specified
if is_closed(review["status"]) and not args.force:
print(
"Change {} status is {}. Skipping the cherry pick.\nUse -f to force this pick.".format(
change, review["status"]
)
)
continue
# Convert the project name to a project path
# - check that the project path exists
if (
review["project"] in project_name_to_data
and review["branch"] in project_name_to_data[review["project"]]
):
project_path = project_name_to_data[review["project"]][review["branch"]]
elif args.path:
project_path = args.path
elif (
review["project"] in project_name_to_data
and len(project_name_to_data[review["project"]]) == 1
):
local_branch = list(project_name_to_data[review["project"]])[0]
project_path = project_name_to_data[review["project"]][local_branch]
print(
'WARNING: Project {0} has a different branch ("{1}" != "{2}")'.format(
project_path, local_branch, review["branch"]
)
)
elif args.ignore_missing:
print(
"WARNING: Skipping {0} since there is no project directory for: {1}\n".format(
review["number"], review["project"]
)
)
continue
else:
sys.stderr.write(
"ERROR: For {0}, could not determine the project path for project {1}\n".format(
review["number"], review["project"]
)
)
sys.exit(1)
item = {
"subject": review["subject"],
"project_path": project_path,
"branch": review["branch"],
"change_id": review["change_id"],
"change_number": review["number"],
"status": review["status"],
"patchset": review["revisions"][review["current_revision"]]["_number"],
"fetch": review["revisions"][review["current_revision"]]["fetch"],
"id": change,
"revision": review["current_revision"],
}
if patchset:
for x in review["revisions"]:
if review["revisions"][x]["_number"] == patchset:
item["fetch"] = review["revisions"][x]["fetch"]
item["id"] = "{0}/{1}".format(change, patchset)
item["patchset"] = patchset
item["revision"] = x
break
else:
if not args.quiet:
print(
"ERROR: The patch set {0}/{1} could not be found, using CURRENT_REVISION instead.".format(
change, patchset
)
)
mergables[project_path].append(item)
# round 1: start branch and drop picked changes
for project_path in mergables:
# If --start-branch is given, create the branch (more than once per path is okay; repo ignores gracefully)
if args.start_branch:
subprocess.run(["repo", "start", args.start_branch[0], project_path])
# Determine the maximum commits to check already picked changes
check_picked_count = args.check_picked
branch_commits_count = int(
subprocess.check_output(
[
"git",
"rev-list",
"--count",
"--max-count",
str(check_picked_count + 1),
"HEAD",
],
cwd=project_path,
text=True,
)
)
if branch_commits_count <= check_picked_count:
check_picked_count = branch_commits_count - 1
picked_change_ids = []
for i in range(check_picked_count):
if not commit_exists(project_path, "HEAD~{0}".format(i)):
continue
output = subprocess.check_output(
["git", "show", "-q", f"HEAD~{i}"], cwd=project_path, text=True
)
output = output.split()
if "Change-Id:" in output:
for j, t in enumerate(reversed(output)):
if t == "Change-Id:":
head_change_id = output[len(output) - j]
picked_change_ids.append(head_change_id.strip())
break
def filter_picked(item):
# Check if change is already picked to HEAD...HEAD~check_picked_count
if item["change_id"] in picked_change_ids:
print(
"Skipping {0} - already picked in {1}".format(
item["id"], project_path
)
)
return False
return True
mergables[project_path] = list(filter(filter_picked, mergables[project_path]))
# round 2: fetch changes in parallel if not pull
if not args.pull:
with ThreadPoolExecutor(max_workers=args.jobs) as e:
for per_path_mergables in mergables.values():
# changes are sorted so loop in reversed order to fetch top commits first
for item in reversed(per_path_mergables):
e.submit(partial(do_git_fetch_pull, args), item)
# round 3: apply changes in parallel for different projects, but sequential
# within each project
with ThreadPoolExecutor(max_workers=args.jobs) as e:
def bulk_pick_change(per_path_mergables):
for item in per_path_mergables:
apply_change(args, item)
for per_path_mergables in mergables.values():
e.submit(bulk_pick_change, per_path_mergables)
def do_git_fetch_pull(args, item):
project_path = item["project_path"]
# commit object already exists, no need to fetch
if not args.pull and commit_exists(project_path, item["revision"]):
return
if "anonymous http" in item["fetch"]:
method = "anonymous http"
else:
method = "ssh"
if args.pull:
cmd = ["git", "pull", "--no-edit"]
else:
cmd = ["git", "fetch"]
if args.quiet:
cmd.append("--quiet")
cmd.extend(["", item["fetch"][method]["ref"]])
# Try fetching from GitHub first if using omnirom gerrit
if is_omnirom_gerrit(args.gerrit):
if args.verbose:
print("Trying to fetch the change from GitHub")
cmd[-2] = "omnirom"
if not args.quiet:
print(cmd)
result = subprocess.call(cmd, cwd=project_path)
# Check if it worked
if result == 0 or commit_exists(project_path, item["revision"]):
return
print("ERROR: git command failed")
# If not using the omnirom gerrit or github failed, fetch from gerrit.
if args.verbose:
if is_omnirom_gerrit(args.gerrit):
print(
"Fetching from GitHub didn't work, trying to fetch the change from Gerrit"
)
else:
print("Fetching from {0}".format(args.gerrit))
cmd[-2] = item["fetch"][method]["url"]
if not args.quiet:
print(cmd)
result = subprocess.call(cmd, cwd=project_path)
if result != 0 and not commit_exists(project_path, item["revision"]):
print("ERROR: git command failed")
sys.exit(result)
def apply_change(args, item):
if not args.quiet:
print("Applying change number {0}...".format(item["id"]))
if is_closed(item["status"]):
print("!! Force-picking a closed change !!\n")
project_path = item["project_path"]
# Print out some useful info
if not args.quiet:
print('--> Subject: "{0}"'.format(item["subject"]))
print("--> Project path: {0}".format(project_path))
print(
"--> Change number: {0} (Patch Set {1})".format(
item["id"], item["patchset"]
)
)
if args.pull:
do_git_fetch_pull(args, item)
else:
# Perform the cherry-pick
if args.quiet:
cmd_out = subprocess.DEVNULL
else:
cmd_out = None
result = subprocess.call(
["git", "cherry-pick", "--ff", item["revision"]],
cwd=project_path,
stdout=cmd_out,
stderr=cmd_out,
)
if result != 0:
result = subprocess.call(
["git", "diff-index", "--quiet", "HEAD", "--"],
cwd=project_path,
stdout=cmd_out,
stderr=cmd_out,
)
if result == 0:
print(
"WARNING: git command resulted with an empty commit, aborting cherry-pick"
)
subprocess.call(
["git", "cherry-pick", "--abort"],
cwd=project_path,
stdout=cmd_out,
stderr=cmd_out,
)
elif args.reset:
print("ERROR: git command failed, aborting cherry-pick")
subprocess.call(
["git", "cherry-pick", "--abort"],
cwd=project_path,
stdout=cmd_out,
stderr=cmd_out,
)
sys.exit(result)
else:
print("ERROR: git command failed")
sys.exit(result)
if not args.quiet:
print("")
if __name__ == "__main__":
main()