summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorOlav Vitters <olav@vitters.nl>2020-04-23 17:33:08 +0200
committerOlav Vitters <olav@vitters.nl>2020-04-23 17:33:08 +0200
commitaeb84f81397f0608da98d12c019fc23c4b038475 (patch)
tree4e30f539da3e640c633bd116cb83fa1446059265
parent50c9a15f69ce67d6cade5c02366cf58d3abf51ed (diff)
downloadmgagnome-aeb84f81397f0608da98d12c019fc23c4b038475.tar
mgagnome-aeb84f81397f0608da98d12c019fc23c4b038475.tar.gz
mgagnome-aeb84f81397f0608da98d12c019fc23c4b038475.tar.bz2
mgagnome-aeb84f81397f0608da98d12c019fc23c4b038475.tar.xz
mgagnome-aeb84f81397f0608da98d12c019fc23c4b038475.zip
pylint: remove various unused variables, avoid len check, etc
-rwxr-xr-xmgagnome81
1 files changed, 36 insertions, 45 deletions
diff --git a/mgagnome b/mgagnome
index d4c19bf..0361850 100755
--- a/mgagnome
+++ b/mgagnome
@@ -1,4 +1,5 @@
#!/usr/bin/python3 -u
+"""Mageia GNOME commands."""
# A lot of the code comes from ftpadmin, see
# https://git.gnome.org/browse/sysadmin-bin/tree/ftpadmin
@@ -165,10 +166,10 @@ def get_safe_max_version(version, module=None):
min_nr = int(majmin[1])
- if min_nr % 2 == 0:
- return "%s.%d" % (majmin[0], min_nr + 1)
- else:
- return "%s.%d" % (majmin[0], min_nr + 2)
+ # Add 2 for stable releases, 1 for unstable
+ min_nr += 1 if min_nr % 2 == 0 else 2
+
+ return "%s.%d" % (majmin[0], min_nr)
def judge_version_increase(version_old, version_new, module=None):
"""Judge quality of version increase:
@@ -278,7 +279,7 @@ def call_editor(filename):
return True
-class urllister(HTMLParser):
+class URLLister(HTMLParser):
def reset(self):
HTMLParser.reset(self)
self.urls = []
@@ -311,10 +312,9 @@ def clean_pkgconfig_prov(prov):
class SpecFileError(Exception):
"""Used for problems in the spec file"""
- pass
-class SpecFile(object):
+class SpecFile():
re_update_version = re.compile(r'^(?P<pre>Version[ \t]*:\s*)(?P<version>.+)(?P<post>\s*)$', re.MULTILINE + re.IGNORECASE)
re_update_release = re.compile(r'^(?P<pre>Release[ \t]*:\s*)(?P<release>%mkrel [0-9.]+)(?P<post>\s*)$', re.MULTILINE + re.IGNORECASE)
re_update_patch = re.compile(r'^(?P<pre>Patch0*?)(?P<nr>[0-9]*)(?P<pre2>[ \t]*:\s*)(?P<patch>.+)(?P<post>\s*)\n', re.MULTILINE + re.IGNORECASE)
@@ -416,7 +416,7 @@ class SpecFile(object):
with open(self.path, "r", encoding="utf-8") as f:
data = f.read()
for reason, change_to, regexp, *extra in re_clean_spec:
- if len(extra):
+ if extra:
should_rebuild = extra[0]
else:
should_rebuild = False
@@ -480,7 +480,7 @@ class SpecFile(object):
data = data.lstrip()
self._changes['SILENT remove variable definition(s) %s' % ", ".join(converted_defines)] = True
- made_changes, data = self._clean_spec_patches(f, made_changes, data)
+ made_changes, data = self._clean_spec_patches(made_changes, data)
# Overwrite file with new version number
if made_changes:
@@ -489,7 +489,7 @@ class SpecFile(object):
return made_changes
- def _clean_spec_patches(self, f, made_changes, data):
+ def _clean_spec_patches(self, made_changes, data):
re_autopatch = re.compile(r'^[ \t]*\%autopatch(?:[ \t]+-p(?P<strip>[0-9]+))?$', re.MULTILINE)
re_patch_header = re.compile('^Patch(?P<nr>[0-9]*)[ \t]*:[ \t]*(?P<filename>[^\n]+)\n', re.MULTILINE + re.IGNORECASE)
@@ -619,12 +619,9 @@ class SpecFile(object):
with open(self.path, "r", encoding="utf-8") as f:
data = f.read()
- len_before = len(data)
-
data, nr = self.re_update_patch.subn(lambda mo: '' if mo.group('nr') in nrs or (mo.group('nr').isdigit() and int(mo.group('nr')) in nrs) else mo.group(0), data)
- # XXX - pretty hacky
- if len(data) == len_before:
+ if not nr:
print("ERROR: Could not remove patch nr %s!" % patchnr, file=sys.stderr)
return False
@@ -746,28 +743,23 @@ class SpecFile(object):
with open(self.path, "r", encoding="utf-8") as f:
data = f.read()
- data_before = data
# Change any "," in buildrequires into multiple lines
data, nr = self.re_update_br_unsplit.subn(lambda mo: ''.join((''.join((mo.group('pre'), mo2.group(0), mo.group('unsplitpost'))) for mo2 in self.re_br_part.finditer(mo.group('unsplit')) if mo.group(0).strip() != '')), data)
- if data_before != data:
+ if nr:
made_changes = True
self._changes['SILENT one line per buildrequirement'] = True
- data_before = data
# Change =< and => operators into <= and >=
- # XXX - pretty ugly
data, nr = self.re_update_br_fix_operator.subn(lambda mo: mo.group(0).replace('=>', '>=').replace('=<', '<') if self.re_update_br.match(mo.group(0).replace('=>', '>=').replace('=<', '<')) else mo.group(0), data)
- if data_before != data:
+ if nr:
made_changes = True
self._changes['SILENT fix operator in buildrequires'] = True
- data_before = data
# Now update buildrequires if any
data, nr = self.re_update_br.subn(lambda mo: ''.join((mo.group('pre'), changes[mo.group('br')], mo.group('post'))) if mo.group('br') in changes else mo.group(0), data)
- # XXX - very hacky because of multiple changes, could miss out on a change
- if data_before != data:
+ if nr:
made_changes = True
self._changes['SILENT %s' % change_description] = True
elif len(changes) != 0:
@@ -822,8 +814,6 @@ class SpecFile(object):
def update_release(self, release, reason, force=False):
"""Update release (usually for rebuilds)"""
- cur_release = self.release
-
data = self._check_can_update(force)
if data is None:
return False
@@ -907,7 +897,7 @@ class SpecFile(object):
return self.check_and_update_patches()
-class Patch(object):
+class Patch():
"""Do things with patches"""
re_dep3 = re.compile(r'^(?:#\s*)?(?P<header>[-A-Za-z0-9]+?):\s*(?P<data>.*)$')
@@ -1048,7 +1038,7 @@ class Patch(object):
return self._svn_author
-class Upstream(object):
+class Upstream():
URL = "https://download.gnome.org/sources/"
limit = None
@@ -1061,7 +1051,7 @@ class Upstream(object):
# Get the files
usock = urlopen.open(self.URL)
- parser = urllister()
+ parser = URLLister()
parser.feed(usock.read().decode('utf-8'))
usock.close()
parser.close()
@@ -1089,7 +1079,7 @@ class Upstream(object):
return cls._cache_versions[module]
-class Downstream(object):
+class Downstream():
re_file = re.compile(r'^(?P<module>.*)[_-](?:(?P<oldversion>([0-9]+[\.])*[0-9]+)-)?(?P<version>([0-9]+[\.\-])*[0-9]+)\.(?P<format>(?:tar\.|diff\.)?[a-z][a-z0-9]*)$')
MEDIA = "Core Release Source"
@@ -1155,7 +1145,7 @@ class Downstream(object):
Sort of works like:
$ urpmq --whatprovides $search_for --provides"""
- if not len(cls._provide_to_alternate):
+ if not cls._provide_to_alternate:
provide_has_multiple_pkgs = set()
_provide_to_pkg = {}
_pkg_to_provide = {}
@@ -1261,7 +1251,7 @@ class Downstream(object):
# Return all packages reflecting the current version
matches = [package for package in packages if packages[package] == version]
- if len(matches):
+ if matches:
return matches
# Return all packages reflecting the version before the current version
@@ -1280,7 +1270,7 @@ class Downstream(object):
# - now really get the right packages
matches = [package for package in packages if packages[package] == latest_version]
- if len(matches):
+ if matches:
return matches
# Give up
@@ -1305,7 +1295,7 @@ def cmd_co_multi(args):
def cmd_co(options, parser):
if options.all:
packages = ((package, package, options) for package in Downstream().packages)
- elif len(options.package):
+ elif options.package:
packages = ((package, package, options) for package in options.package)
else:
packages = ((l[0], "%s => %s" % (l[0], l[1]), options) for l in sorted(join_streams(auto_update=False)))
@@ -1425,7 +1415,7 @@ def cmd_cleanup(options, parser):
c2 = pysvn.Client()
stats = [stat for stat in c2.status(os.path.join(root, path, 'SOURCES'), depth=pysvn.depth.immediates) if stat.text_status == pysvn.wc_status_kind.unversioned and os.path.basename(stat.path) not in binaries]
- if len(stats):
+ if stats:
print(path)
print(", ".join(os.path.basename(stat.path) for stat in stats))
print(stats)
@@ -1625,7 +1615,7 @@ def cmd_clean_spec_multi(args):
change_to = None
if len(provides_alt) == 1:
change_to = provides_alt[0]
- elif len(provides_alt) and 'extra' in keys:
+ elif provides_alt and 'extra' in keys:
# Determine base require (e.g. gtk+3.0-devel --> gtk+3.0)
basereqs = keys['basereqs'](req)
@@ -1634,7 +1624,7 @@ def cmd_clean_spec_multi(args):
if 'versions_from_basereq' in keys:
# Determine if the basereq has a version at the end (e.g. gtk+3.0 --> 3.0)
versions.update(keys['versions_from_basereq'](basereqs))
- if len(versions) and 'basereq_no_version' in keys:
+ if versions and 'basereq_no_version' in keys:
basereqs.extend(keys['basereq_no_version'](basereqs))
# Make it unique again, but keep the order
#
@@ -1644,7 +1634,7 @@ def cmd_clean_spec_multi(args):
if 'versions_basereq_extra' in keys:
versions.update(keys['versions_basereq_extra'](versions))
- if not len(versions):
+ if not versions:
# In case no versions were retrieved from the basereq,
# match with any version found from the alternative
# provides (heuristic matching)
@@ -1667,7 +1657,7 @@ def cmd_clean_spec_multi(args):
change_to = check
break
- if change_to is None and len(provides_alt):
+ if change_to is None and provides_alt:
provides_alt_no_versions = []
for prov in provides_alt:
if re_prov_get_version.fullmatch(prov) is None:
@@ -1677,7 +1667,7 @@ def cmd_clean_spec_multi(args):
change_to = provides_alt_no_versions[0]
- if len(provides_alt):
+ if provides_alt:
if change_to is None: no_change[req] = (provides_alt, check_for)
else:
no_alt.add(req)
@@ -1719,6 +1709,8 @@ def cmd_clean_spec_multi(args):
print(s.changes)
s.ensure_no_local_changes(force=True)
+ return made_changes
+
def cmd_check_spec_multi(args):
options, package = args
cwd = Downstream.package_path(package)
@@ -1742,14 +1734,13 @@ def cmd_check_spec(options, parser):
if options.all:
packages = Downstream().packages
else:
- packages = options.package if len(options.package) else (l[0] for l in join_streams())
+ packages = options.package if options.package else (l[0] for l in join_streams())
if options.debug:
for package in packages:
cmd_check_spec_multi((options, package))
else:
- import os
workers = os.cpu_count() or 4
with concurrent.futures.ThreadPoolExecutor(max_workers=workers) as executor:
executor.map(cmd_check_spec_multi, ((options, package) for package in packages))
@@ -1760,14 +1751,13 @@ def cmd_clean_spec(options, parser):
if options.all:
packages = Downstream().packages
else:
- packages = options.package if len(options.package) else (l[0] for l in join_streams())
+ packages = options.package if options.package else (l[0] for l in join_streams())
if options.debug:
for package in packages:
cmd_clean_spec_multi((options, package))
else:
- import os
workers = os.cpu_count() or 4
# Hack: warm alternative provides cache
if options.convert_br:
@@ -1884,9 +1874,9 @@ def cmd_package_new_version(options, parser):
# Check hash, if given
if options.hexdigest is not None:
sources = [name for name, value in s.sources.items() if '://' in value[0]]
- if not len(sources):
+ if not sources:
print("ERROR: Cannot determine source file (for hash check)!", file=sys.stderr)
- sys.stderr(1)
+ sys.exit(1)
# If there are multiple sources, try to see if there is a preferred name
# --> needed for metacity hash check (multiple tarball sources)
@@ -1917,7 +1907,8 @@ def cmd_package_new_version(options, parser):
# retry submission various times, could be that some dependencies are being built at the same time
@retry(subprocess.CalledProcessError, tries=10, delay=300, backoff=1.5)
- def _submit(): subprocess.check_call(cmd, cwd=cwd)
+ def _submit():
+ subprocess.check_call(cmd, cwd=cwd)
_submit()
except subprocess.CalledProcessError:
sys.exit(1)