mirror of
https://git.busybox.net/buildroot.git
synced 2024-11-24 05:53:30 +08:00
de03312db0
Start counting the line numbers in 1 instead of 0, in case an error must be printed. Both the error about a developer entry with no file entry and the error about a file entry with no developer entry actually belong to the non-empty line previous the one being analysed, so in these cases print the line number from the line before. Also count empty and comment lines, so a developer fixing the file can jump to the correct line (or the nearest one). At same time standardize the messages, printing the line number also in the case of a warning for a file that is not in the tree anymore. Cc: Thomas Petazzoni <thomas.petazzoni@bootlin.com> Signed-off-by: Ricardo Martincoski <ricardo.martincoski@gmail.com> Signed-off-by: Thomas Petazzoni <thomas.petazzoni@bootlin.com>
294 lines
9.4 KiB
Python
294 lines
9.4 KiB
Python
from io import open
|
|
import os
|
|
import re
|
|
import glob
|
|
import subprocess
|
|
import sys
|
|
import unittest
|
|
|
|
brpath = os.path.normpath(os.path.join(os.path.dirname(__file__), ".."))
|
|
|
|
#
|
|
# Patch parsing functions
|
|
#
|
|
|
|
FIND_INFRA_IN_PATCH = re.compile(r"^\+\$\(eval \$\((host-)?([^-]*)-package\)\)$")
|
|
|
|
|
|
def analyze_patch(patch):
|
|
"""Parse one patch and return the list of files modified, added or
|
|
removed by the patch."""
|
|
files = set()
|
|
infras = set()
|
|
for line in patch:
|
|
# If the patch is adding a package, find which infra it is
|
|
m = FIND_INFRA_IN_PATCH.match(line)
|
|
if m:
|
|
infras.add(m.group(2))
|
|
if not line.startswith("+++ "):
|
|
continue
|
|
line.strip()
|
|
fname = line[line.find("/") + 1:].strip()
|
|
if fname == "dev/null":
|
|
continue
|
|
files.add(fname)
|
|
return (files, infras)
|
|
|
|
|
|
FIND_INFRA_IN_MK = re.compile(r"^\$\(eval \$\((host-)?([^-]*)-package\)\)$")
|
|
|
|
|
|
def fname_get_package_infra(fname):
|
|
"""Checks whether the file name passed as argument is a Buildroot .mk
|
|
file describing a package, and find the infrastructure it's using."""
|
|
if not fname.endswith(".mk"):
|
|
return None
|
|
|
|
if not os.path.exists(fname):
|
|
return None
|
|
|
|
with open(fname, "r") as f:
|
|
for line in f:
|
|
line = line.strip()
|
|
m = FIND_INFRA_IN_MK.match(line)
|
|
if m:
|
|
return m.group(2)
|
|
return None
|
|
|
|
|
|
def analyze_patches(patches):
|
|
"""Parse a list of patches and returns the list of files modified,
|
|
added or removed by the patches, as well as the list of package
|
|
infrastructures used by those patches (if any)"""
|
|
allfiles = set()
|
|
allinfras = set()
|
|
for patch in patches:
|
|
(files, infras) = analyze_patch(patch)
|
|
allfiles = allfiles | files
|
|
allinfras = allinfras | infras
|
|
return (allfiles, allinfras)
|
|
|
|
|
|
#
|
|
# Unit-test parsing functions
|
|
#
|
|
|
|
def get_all_test_cases(suite):
|
|
"""Generate all test-cases from a given test-suite.
|
|
:return: (test.module, test.name)"""
|
|
if issubclass(type(suite), unittest.TestSuite):
|
|
for test in suite:
|
|
for res in get_all_test_cases(test):
|
|
yield res
|
|
else:
|
|
yield (suite.__module__, suite.__class__.__name__)
|
|
|
|
|
|
def list_unittests():
|
|
"""Use the unittest module to retreive all test cases from a given
|
|
directory"""
|
|
loader = unittest.TestLoader()
|
|
suite = loader.discover(os.path.join(brpath, "support", "testing"))
|
|
tests = {}
|
|
for module, test in get_all_test_cases(suite):
|
|
module_path = os.path.join("support", "testing", *module.split('.'))
|
|
tests.setdefault(module_path, []).append('%s.%s' % (module, test))
|
|
return tests
|
|
|
|
|
|
unittests = {}
|
|
|
|
|
|
#
|
|
# DEVELOPERS file parsing functions
|
|
#
|
|
|
|
class Developer:
|
|
def __init__(self, name, files):
|
|
self.name = name
|
|
self.files = files
|
|
self.packages = parse_developer_packages(files)
|
|
self.architectures = parse_developer_architectures(files)
|
|
self.infras = parse_developer_infras(files)
|
|
self.runtime_tests = parse_developer_runtime_tests(files)
|
|
self.defconfigs = parse_developer_defconfigs(files)
|
|
|
|
def hasfile(self, f):
|
|
for fs in self.files:
|
|
if f.startswith(fs):
|
|
return True
|
|
return False
|
|
|
|
def __repr__(self):
|
|
name = '\'' + self.name.split(' <')[0][:20] + '\''
|
|
things = []
|
|
if len(self.files):
|
|
things.append('{} files'.format(len(self.files)))
|
|
if len(self.packages):
|
|
things.append('{} pkgs'.format(len(self.packages)))
|
|
if len(self.architectures):
|
|
things.append('{} archs'.format(len(self.architectures)))
|
|
if len(self.infras):
|
|
things.append('{} infras'.format(len(self.infras)))
|
|
if len(self.runtime_tests):
|
|
things.append('{} tests'.format(len(self.runtime_tests)))
|
|
if len(self.defconfigs):
|
|
things.append('{} defconfigs'.format(len(self.defconfigs)))
|
|
if things:
|
|
return 'Developer <{} ({})>'.format(name, ', '.join(things))
|
|
else:
|
|
return 'Developer <' + name + '>'
|
|
|
|
|
|
def parse_developer_packages(fnames):
|
|
"""Given a list of file patterns, travel through the Buildroot source
|
|
tree to find which packages are implemented by those file
|
|
patterns, and return a list of those packages."""
|
|
packages = set()
|
|
for fname in fnames:
|
|
for root, dirs, files in os.walk(os.path.join(brpath, fname)):
|
|
for f in files:
|
|
path = os.path.join(root, f)
|
|
if fname_get_package_infra(path):
|
|
pkg = os.path.splitext(f)[0]
|
|
packages.add(pkg)
|
|
return packages
|
|
|
|
|
|
def parse_arches_from_config_in(fname):
|
|
"""Given a path to an arch/Config.in.* file, parse it to get the list
|
|
of BR2_ARCH values for this architecture."""
|
|
arches = set()
|
|
with open(fname, "r") as f:
|
|
parsing_arches = False
|
|
for line in f:
|
|
line = line.strip()
|
|
if line == "config BR2_ARCH":
|
|
parsing_arches = True
|
|
continue
|
|
if parsing_arches:
|
|
m = re.match(r"^\s*default \"([^\"]*)\".*", line)
|
|
if m:
|
|
arches.add(m.group(1))
|
|
else:
|
|
parsing_arches = False
|
|
return arches
|
|
|
|
|
|
def parse_developer_architectures(fnames):
|
|
"""Given a list of file names, find the ones starting by
|
|
'arch/Config.in.', and use that to determine the architecture a
|
|
developer is working on."""
|
|
arches = set()
|
|
for fname in fnames:
|
|
if not re.match(r"^.*/arch/Config\.in\..*$", fname):
|
|
continue
|
|
arches = arches | parse_arches_from_config_in(fname)
|
|
return arches
|
|
|
|
|
|
def parse_developer_infras(fnames):
|
|
infras = set()
|
|
for fname in fnames:
|
|
m = re.match(r"^package/pkg-([^.]*).mk$", fname)
|
|
if m:
|
|
infras.add(m.group(1))
|
|
return infras
|
|
|
|
|
|
def parse_developer_defconfigs(fnames):
|
|
"""Given a list of file names, returns the config names
|
|
corresponding to defconfigs."""
|
|
return {os.path.basename(fname[:-10])
|
|
for fname in fnames
|
|
if fname.endswith('_defconfig')}
|
|
|
|
|
|
def parse_developer_runtime_tests(fnames):
|
|
"""Given a list of file names, returns the runtime tests
|
|
corresponding to the file."""
|
|
all_files = []
|
|
# List all files recursively
|
|
for fname in fnames:
|
|
if os.path.isdir(fname):
|
|
for root, _dirs, files in os.walk(os.path.join(brpath, fname)):
|
|
all_files += [os.path.join(root, f) for f in files]
|
|
else:
|
|
all_files.append(fname)
|
|
|
|
# Get all runtime tests
|
|
runtimes = set()
|
|
for f in all_files:
|
|
name = os.path.splitext(f)[0]
|
|
if name in unittests:
|
|
runtimes |= set(unittests[name])
|
|
return runtimes
|
|
|
|
|
|
def parse_developers(filename=None):
|
|
"""Parse the DEVELOPERS file and return a list of Developer objects."""
|
|
developers = []
|
|
linen = 0
|
|
global unittests
|
|
unittests = list_unittests()
|
|
developers_fname = filename or os.path.join(brpath, 'DEVELOPERS')
|
|
with open(developers_fname, mode='r', encoding='utf_8') as f:
|
|
files = []
|
|
name = None
|
|
for line in f:
|
|
linen += 1
|
|
line = line.strip()
|
|
if line.startswith("#"):
|
|
continue
|
|
elif line.startswith("N:"):
|
|
if name is not None or len(files) != 0:
|
|
print("Syntax error in DEVELOPERS file, line %d" % (linen - 1),
|
|
file=sys.stderr)
|
|
return None
|
|
name = line[2:].strip()
|
|
elif line.startswith("F:"):
|
|
fname = line[2:].strip()
|
|
dev_files = glob.glob(os.path.join(brpath, fname))
|
|
if len(dev_files) == 0:
|
|
print("WARNING: '%s' doesn't match any file, line %d" % (fname, linen),
|
|
file=sys.stderr)
|
|
for f in dev_files:
|
|
dev_file = os.path.relpath(f, brpath)
|
|
dev_file = dev_file.replace(os.sep, '/') # force unix sep
|
|
if f[-1] == '/': # relpath removes the trailing /
|
|
dev_file = dev_file + '/'
|
|
files.append(dev_file)
|
|
elif line == "":
|
|
if not name:
|
|
continue
|
|
developers.append(Developer(name, files))
|
|
files = []
|
|
name = None
|
|
else:
|
|
print("Syntax error in DEVELOPERS file, line %d: '%s'" % (linen, line),
|
|
file=sys.stderr)
|
|
return None
|
|
# handle last developer
|
|
if name is not None:
|
|
developers.append(Developer(name, files))
|
|
return developers
|
|
|
|
|
|
def check_developers(developers, basepath=None):
|
|
"""Look at the list of files versioned in Buildroot, and returns the
|
|
list of files that are not handled by any developer"""
|
|
if basepath is None:
|
|
basepath = os.getcwd()
|
|
cmd = ["git", "--git-dir", os.path.join(basepath, ".git"), "ls-files"]
|
|
files = subprocess.check_output(cmd).decode(sys.stdout.encoding).strip().split("\n")
|
|
unhandled_files = []
|
|
for f in files:
|
|
handled = False
|
|
for d in developers:
|
|
if d.hasfile(f):
|
|
handled = True
|
|
break
|
|
if not handled:
|
|
unhandled_files.append(f)
|
|
return unhandled_files
|