61ddb1aa07
* build: bump pylint to 2.17 Xref: https://chromium-review.googlesource.com/c/chromium/tools/depot_tools/+/5062345 * fix pylint consider-using-f-string warnings pt 1: use flynt for automated fixes * fix pylint consider-using-f-string warnings pt 2: manual fixes * fix pylint consider-using-with warnings * fix pylint line-too-long warnings * fix pylint unspecified-encoding warnings * fix py lint consider-using-generator warning * fixup! fix pylint unspecified-encoding warnings * fix pylint line-too-long warnings
289 lines
8.6 KiB
Python
289 lines
8.6 KiB
Python
#!/usr/bin/env python3
|
|
|
|
"""Git helper functions.
|
|
|
|
Everything here should be project agnostic: it shouldn't rely on project's
|
|
structure, or make assumptions about the passed arguments or calls' outcomes.
|
|
"""
|
|
|
|
import io
|
|
import os
|
|
import posixpath
|
|
import re
|
|
import subprocess
|
|
import sys
|
|
|
|
SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__))
|
|
sys.path.append(SCRIPT_DIR)
|
|
|
|
from patches import PATCH_FILENAME_PREFIX, is_patch_location_line
|
|
|
|
UPSTREAM_HEAD='refs/patches/upstream-head'
|
|
|
|
def is_repo_root(path):
|
|
path_exists = os.path.exists(path)
|
|
if not path_exists:
|
|
return False
|
|
|
|
git_folder_path = os.path.join(path, '.git')
|
|
git_folder_exists = os.path.exists(git_folder_path)
|
|
|
|
return git_folder_exists
|
|
|
|
|
|
def get_repo_root(path):
|
|
"""Finds a closest ancestor folder which is a repo root."""
|
|
norm_path = os.path.normpath(path)
|
|
norm_path_exists = os.path.exists(norm_path)
|
|
if not norm_path_exists:
|
|
return None
|
|
|
|
if is_repo_root(norm_path):
|
|
return norm_path
|
|
|
|
parent_path = os.path.dirname(norm_path)
|
|
|
|
# Check if we're in the root folder already.
|
|
if parent_path == norm_path:
|
|
return None
|
|
|
|
return get_repo_root(parent_path)
|
|
|
|
|
|
def am(repo, patch_data, threeway=False, directory=None, exclude=None,
|
|
committer_name=None, committer_email=None, keep_cr=True):
|
|
args = []
|
|
if threeway:
|
|
args += ['--3way']
|
|
if directory is not None:
|
|
args += ['--directory', directory]
|
|
if exclude is not None:
|
|
for path_pattern in exclude:
|
|
args += ['--exclude', path_pattern]
|
|
if keep_cr is True:
|
|
# Keep the CR of CRLF in case any patches target files with Windows line
|
|
# endings.
|
|
args += ['--keep-cr']
|
|
|
|
root_args = ['-C', repo]
|
|
if committer_name is not None:
|
|
root_args += ['-c', 'user.name=' + committer_name]
|
|
if committer_email is not None:
|
|
root_args += ['-c', 'user.email=' + committer_email]
|
|
root_args += ['-c', 'commit.gpgsign=false']
|
|
command = ['git'] + root_args + ['am'] + args
|
|
with subprocess.Popen(command, stdin=subprocess.PIPE) as proc:
|
|
proc.communicate(patch_data.encode('utf-8'))
|
|
if proc.returncode != 0:
|
|
raise RuntimeError(f"Command {command} returned {proc.returncode}")
|
|
|
|
|
|
def import_patches(repo, ref=UPSTREAM_HEAD, **kwargs):
|
|
"""same as am(), but we save the upstream HEAD so we can refer to it when we
|
|
later export patches"""
|
|
update_ref(repo=repo, ref=ref, newvalue='HEAD')
|
|
am(repo=repo, **kwargs)
|
|
|
|
|
|
def update_ref(repo, ref, newvalue):
|
|
args = ['git', '-C', repo, 'update-ref', ref, newvalue]
|
|
|
|
return subprocess.check_call(args)
|
|
|
|
|
|
def get_commit_for_ref(repo, ref):
|
|
args = ['git', '-C', repo, 'rev-parse', '--verify', ref]
|
|
return subprocess.check_output(args).decode('utf-8').strip()
|
|
|
|
def get_commit_count(repo, commit_range):
|
|
args = ['git', '-C', repo, 'rev-list', '--count', commit_range]
|
|
return int(subprocess.check_output(args).decode('utf-8').strip())
|
|
|
|
def guess_base_commit(repo, ref):
|
|
"""Guess which commit the patches might be based on"""
|
|
try:
|
|
upstream_head = get_commit_for_ref(repo, ref)
|
|
num_commits = get_commit_count(repo, upstream_head + '..')
|
|
return [upstream_head, num_commits]
|
|
except subprocess.CalledProcessError:
|
|
args = [
|
|
'git',
|
|
'-C',
|
|
repo,
|
|
'describe',
|
|
'--tags',
|
|
]
|
|
return subprocess.check_output(args).decode('utf-8').rsplit('-', 2)[0:2]
|
|
|
|
|
|
def format_patch(repo, since):
|
|
args = [
|
|
'git',
|
|
'-C',
|
|
repo,
|
|
'-c',
|
|
'core.attributesfile='
|
|
+ os.path.join(
|
|
os.path.dirname(os.path.realpath(__file__)),
|
|
'electron.gitattributes',
|
|
),
|
|
# Ensure it is not possible to match anything
|
|
# Disabled for now as we have consistent chunk headers
|
|
# '-c',
|
|
# 'diff.electron.xfuncname=$^',
|
|
'format-patch',
|
|
'--keep-subject',
|
|
'--no-stat',
|
|
'--stdout',
|
|
|
|
# Per RFC 3676 the signature is separated from the body by a line with
|
|
# '-- ' on it. If the signature option is omitted the signature defaults
|
|
# to the Git version number.
|
|
'--no-signature',
|
|
|
|
# The name of the parent commit object isn't useful information in this
|
|
# context, so zero it out to avoid needless patch-file churn.
|
|
'--zero-commit',
|
|
|
|
# Some versions of git print out different numbers of characters in the
|
|
# 'index' line of patches, so pass --full-index to get consistent
|
|
# behaviour.
|
|
'--full-index',
|
|
since
|
|
]
|
|
return subprocess.check_output(args).decode('utf-8')
|
|
|
|
|
|
def split_patches(patch_data):
|
|
"""Split a concatenated series of patches into N separate patches"""
|
|
patches = []
|
|
patch_start = re.compile('^From [0-9a-f]+ ')
|
|
# Keep line endings in case any patches target files with CRLF.
|
|
keep_line_endings = True
|
|
for line in patch_data.splitlines(keep_line_endings):
|
|
if patch_start.match(line):
|
|
patches.append([])
|
|
patches[-1].append(line)
|
|
return patches
|
|
|
|
def filter_patches(patches, key):
|
|
"""Return patches that include the specified key"""
|
|
if key is None:
|
|
return patches
|
|
matches = []
|
|
for patch in patches:
|
|
if any(key in line for line in patch):
|
|
matches.append(patch)
|
|
continue
|
|
return matches
|
|
|
|
def munge_subject_to_filename(subject):
|
|
"""Derive a suitable filename from a commit's subject"""
|
|
if subject.endswith('.patch'):
|
|
subject = subject[:-6]
|
|
return re.sub(r'[^A-Za-z0-9-]+', '_', subject).strip('_').lower() + '.patch'
|
|
|
|
|
|
def get_file_name(patch):
|
|
"""Return the name of the file to which the patch should be written"""
|
|
file_name = None
|
|
for line in patch:
|
|
if line.startswith(PATCH_FILENAME_PREFIX):
|
|
file_name = line[len(PATCH_FILENAME_PREFIX):]
|
|
break
|
|
# If no patch-filename header, munge the subject.
|
|
if not file_name:
|
|
for line in patch:
|
|
if line.startswith('Subject: '):
|
|
file_name = munge_subject_to_filename(line[len('Subject: '):])
|
|
break
|
|
return file_name.rstrip('\n')
|
|
|
|
|
|
def join_patch(patch):
|
|
"""Joins and formats patch contents"""
|
|
return ''.join(remove_patch_location(patch)).rstrip('\n') + '\n'
|
|
|
|
|
|
def remove_patch_location(patch):
|
|
"""Strip out the patch location lines from a patch's message body"""
|
|
force_keep_next_line = False
|
|
n = len(patch)
|
|
for i, l in enumerate(patch):
|
|
skip_line = is_patch_location_line(l)
|
|
skip_next = i < n - 1 and is_patch_location_line(patch[i + 1])
|
|
if not force_keep_next_line and (
|
|
skip_line or (skip_next and len(l.rstrip()) == 0)
|
|
):
|
|
pass # drop this line
|
|
else:
|
|
yield l
|
|
force_keep_next_line = l.startswith('Subject: ')
|
|
|
|
|
|
def export_patches(repo, out_dir,
|
|
patch_range=None, ref=UPSTREAM_HEAD,
|
|
dry_run=False, grep=None):
|
|
if not os.path.exists(repo):
|
|
sys.stderr.write(
|
|
f"Skipping patches in {repo} because it does not exist.\n"
|
|
)
|
|
return
|
|
if patch_range is None:
|
|
patch_range, n_patches = guess_base_commit(repo, ref)
|
|
msg = f"Exporting {n_patches} patches in {repo} since {patch_range[0:7]}\n"
|
|
sys.stderr.write(msg)
|
|
patch_data = format_patch(repo, patch_range)
|
|
patches = split_patches(patch_data)
|
|
if grep:
|
|
olen = len(patches)
|
|
patches = filter_patches(patches, grep)
|
|
sys.stderr.write(f"Exporting {len(patches)} of {olen} patches\n")
|
|
|
|
try:
|
|
os.mkdir(out_dir)
|
|
except OSError:
|
|
pass
|
|
|
|
if dry_run:
|
|
# If we're doing a dry run, iterate through each patch and see if the newly
|
|
# exported patch differs from what exists. Report number of mismatched
|
|
# patches and fail if there's more than one.
|
|
bad_patches = []
|
|
for patch in patches:
|
|
filename = get_file_name(patch)
|
|
filepath = posixpath.join(out_dir, filename)
|
|
with io.open(filepath, 'rb') as inp:
|
|
existing_patch = str(inp.read(), 'utf-8')
|
|
formatted_patch = join_patch(patch)
|
|
if formatted_patch != existing_patch:
|
|
bad_patches.append(filename)
|
|
if len(bad_patches) > 0:
|
|
sys.stderr.write(
|
|
"Patches in {} not up to date: {} patches need update\n-- {}\n".format(
|
|
out_dir, len(bad_patches), "\n-- ".join(bad_patches)
|
|
)
|
|
)
|
|
sys.exit(1)
|
|
else:
|
|
# Remove old patches so that deleted commits are correctly reflected in the
|
|
# patch files (as a removed file)
|
|
for p in os.listdir(out_dir):
|
|
if p.endswith('.patch'):
|
|
os.remove(posixpath.join(out_dir, p))
|
|
with io.open(
|
|
posixpath.join(out_dir, '.patches'),
|
|
'w',
|
|
newline='\n',
|
|
encoding='utf-8',
|
|
) as pl:
|
|
for patch in patches:
|
|
filename = get_file_name(patch)
|
|
file_path = posixpath.join(out_dir, filename)
|
|
formatted_patch = join_patch(patch)
|
|
# Write in binary mode to retain mixed line endings on write.
|
|
with io.open(
|
|
file_path, 'wb'
|
|
) as f:
|
|
f.write(formatted_patch.encode('utf-8'))
|
|
pl.write(filename + '\n')
|