servo: Merge #14715 - Tidy: Check Cargo.lock for packages with same version and different sources (from UK992:tidy-check-lock); r=SimonSapin
<!-- Please describe your changes on the following line: --> r? @Wafflespeanut cc @SimonSapin --- <!-- Thank you for contributing to Servo! Please replace each `[ ]` by `[X]` when the step is complete, and replace `__` with appropriate data: --> - [X] `./mach build -d` does not report any errors - [X] `./mach test-tidy` does not report any errors - [X] These changes fix #14695 <!-- Either: --> - [X] There are tests for these changes <!-- Pull requests that do not address these steps are welcome, but they will require additional verification as part of the review process. --> Source-Repo: https://github.com/servo/servo Source-Revision: 37a5e29147f0dc489888377d6f7bb53282dc04f9
This commit is contained in:
@@ -33,10 +33,10 @@ config = {
|
||||
"lint-scripts": [],
|
||||
"ignore": {
|
||||
"files": [
|
||||
"./.", # ignore hidden files
|
||||
os.path.join(".", "."), # ignore hidden files
|
||||
],
|
||||
"directories": [
|
||||
"./.", # ignore hidden directories
|
||||
os.path.join(".", "."), # ignore hidden directories
|
||||
],
|
||||
"packages": [],
|
||||
},
|
||||
@@ -90,6 +90,10 @@ def is_iter_empty(iterator):
|
||||
return False, iterator
|
||||
|
||||
|
||||
def normilize_paths(paths):
|
||||
return [os.path.join(*path.split('/')) for path in paths]
|
||||
|
||||
|
||||
# A simple wrapper for iterators to show progress
|
||||
# (Note that it's inefficient for giant iterators, since it iterates once to get the upper bound)
|
||||
def progress_wrapper(iterator):
|
||||
@@ -123,11 +127,9 @@ class FileList(object):
|
||||
args = ["git", "log", "-n1", "--merges", "--format=%H"]
|
||||
last_merge = subprocess.check_output(args).strip()
|
||||
args = ["git", "diff", "--name-only", last_merge, self.directory]
|
||||
file_list = subprocess.check_output(args)
|
||||
file_list = normilize_paths(subprocess.check_output(args).splitlines())
|
||||
|
||||
for f in file_list.splitlines():
|
||||
if sys.platform == 'win32':
|
||||
os.path.join(*f.split('/'))
|
||||
for f in file_list:
|
||||
if not any(os.path.join('.', os.path.dirname(f)).startswith(path) for path in self.excluded):
|
||||
yield os.path.join('.', f)
|
||||
|
||||
@@ -299,61 +301,42 @@ def check_flake8(file_name, contents):
|
||||
|
||||
|
||||
def check_lock(file_name, contents):
|
||||
def find_reverse_dependencies(dependency, version, content):
|
||||
dependency_prefix = "{} {}".format(dependency, version)
|
||||
def find_reverse_dependencies(name, content):
|
||||
for package in itertools.chain([content["root"]], content["package"]):
|
||||
for dependency in package.get("dependencies", []):
|
||||
if dependency.startswith(dependency_prefix):
|
||||
yield package["name"]
|
||||
if dependency.startswith("{} ".format(name)):
|
||||
yield package["name"], dependency
|
||||
|
||||
if not file_name.endswith(".lock"):
|
||||
raise StopIteration
|
||||
|
||||
# package names to be neglected (as named by cargo)
|
||||
# Package names to be neglected (as named by cargo)
|
||||
exceptions = config["ignore"]["packages"]
|
||||
|
||||
# toml.py has a bug(?) that we trip up in [metadata] sections;
|
||||
# see https://github.com/uiri/toml/issues/61
|
||||
# This should only affect a very few lines (that have embedded ?branch=...),
|
||||
# and most of them won't be in the repo
|
||||
try:
|
||||
content = toml.loads(contents)
|
||||
except:
|
||||
print "WARNING!"
|
||||
print "WARNING! toml parsing failed for Cargo.lock, but ignoring..."
|
||||
print "WARNING!"
|
||||
raise StopIteration
|
||||
content = toml.loads(contents)
|
||||
|
||||
packages = {}
|
||||
packages_by_name = {}
|
||||
for package in content.get("package", []):
|
||||
packages.setdefault(package["name"], []).append(package["version"])
|
||||
source = package.get("source", "")
|
||||
if source == r"registry+https://github.com/rust-lang/crates.io-index":
|
||||
source = "crates.io"
|
||||
packages_by_name.setdefault(package["name"], []).append((package["version"], source))
|
||||
|
||||
for (name, versions) in packages.iteritems():
|
||||
if name in exceptions or len(versions) <= 1:
|
||||
for (name, packages) in packages_by_name.iteritems():
|
||||
if name in exceptions or len(packages) <= 1:
|
||||
continue
|
||||
|
||||
highest = max(versions)
|
||||
for version in versions:
|
||||
if version != highest:
|
||||
reverse_dependencies = "\n".join(
|
||||
"\t\t{}".format(n)
|
||||
for n in find_reverse_dependencies(name, version, content)
|
||||
)
|
||||
substitutions = {
|
||||
"package": name,
|
||||
"old_version": version,
|
||||
"new_version": highest,
|
||||
"reverse_dependencies": reverse_dependencies
|
||||
}
|
||||
message = """
|
||||
duplicate versions for package "{package}"
|
||||
\t\033[93mfound dependency on version {old_version}\033[0m
|
||||
\t\033[91mbut highest version is {new_version}\033[0m
|
||||
\t\033[93mtry upgrading with\033[0m \033[96m./mach cargo-update -p {package}:{old_version}\033[0m
|
||||
\tThe following packages depend on version {old_version}:
|
||||
{reverse_dependencies}
|
||||
""".format(**substitutions).strip()
|
||||
yield (1, message)
|
||||
message = "duplicate versions for package `{}`".format(name)
|
||||
packages.sort()
|
||||
packages_dependencies = list(find_reverse_dependencies(name, content))
|
||||
for version, source in packages:
|
||||
short_source = source.split("#")[0].replace("git+", "")
|
||||
message += "\n\t\033[93mThe following packages depend on version {} from '{}':\033[0m" \
|
||||
.format(version, short_source)
|
||||
for name, dependency in packages_dependencies:
|
||||
if version in dependency and short_source in dependency:
|
||||
message += "\n\t\t" + name
|
||||
yield (1, message)
|
||||
|
||||
|
||||
def check_toml(file_name, lines):
|
||||
@@ -862,23 +845,17 @@ def parse_config(content):
|
||||
config_file = toml.loads(content)
|
||||
exclude = config_file.get("ignore", {})
|
||||
# Add list of ignored directories to config
|
||||
config["ignore"]["directories"] += exclude.get("directories", [])
|
||||
config["ignore"]["directories"] += normilize_paths(exclude.get("directories", []))
|
||||
# Add list of ignored files to config
|
||||
config["ignore"]["files"] += exclude.get("files", [])
|
||||
config["ignore"]["files"] += normilize_paths(exclude.get("files", []))
|
||||
# Add list of ignored packages to config
|
||||
config["ignore"]["packages"] = exclude.get("packages", [])
|
||||
# Fix the paths (OS-dependent)
|
||||
config['ignore']['files'] = map(lambda path: os.path.join(*path.split('/')),
|
||||
config['ignore']['files'])
|
||||
config['ignore']['directories'] = map(lambda path: os.path.join(*path.split('/')),
|
||||
config['ignore']['directories'])
|
||||
|
||||
# Add dict of dir, list of expected ext to config
|
||||
dirs_to_check = config_file.get("check_ext", {})
|
||||
# Fix the paths (OS-dependent)
|
||||
for path, exts in dirs_to_check.items():
|
||||
fixed_path = os.path.join(*path.split('/'))
|
||||
config['check_ext'][fixed_path] = exts
|
||||
config['check_ext'][normilize_paths([path])[0]] = exts
|
||||
|
||||
# Override default configs
|
||||
user_configs = config_file.get("configs", [])
|
||||
|
||||
Reference in New Issue
Block a user