1065 lines
37 KiB
Executable File

#!/usr/bin/env python
"""Check the file in a Python source package for completeness.
This script works by building a source distribution archive (by running sdist), then checking the file list in the archive against the
file list in version control (Subversion, Git, Mercurial, Bazaar are
Since the first check can fail to catch missing entries when
you've got the right setuptools version control system support plugins
installed, the script copies all the versioned files into a temporary
directory and building the source distribution again. This also avoids issues
with stale egg-info/SOURCES.txt files that may cause files not mentioned in to be included nevertheless.
from __future__ import print_function
import argparse
import codecs
import fnmatch
import locale
import os
import posixpath
import re
import shutil
import stat
import subprocess
import sys
import tarfile
import tempfile
import unicodedata
import zipfile
from distutils.filelist import glob_to_re
from contextlib import contextmanager, closing
from distutils.text_file import TextFile
from xml.etree import cElementTree as ET
import ConfigParser
except ImportError:
# Python 3.x
import configparser as ConfigParser
import toml
__version__ = '0.41'
__author__ = 'Marius Gedminas <>'
__licence__ = 'MIT'
__url__ = ''
class Failure(Exception):
"""An expected failure (as opposed to a bug in this script)."""
# User interface
_to_be_continued = False
def _check_tbc():
global _to_be_continued
if _to_be_continued:
_to_be_continued = False
def info(message):
def info_begin(message):
if not VERBOSE:
global _to_be_continued
_to_be_continued = True
def info_continue(message):
if not VERBOSE:
global _to_be_continued
_to_be_continued = True
def info_end(message):
if not VERBOSE:
global _to_be_continued
_to_be_continued = False
def error(message):
print(message, file=sys.stderr)
def warning(message):
print(message, file=sys.stderr)
def format_list(list_of_strings):
return "\n".join(" " + s for s in list_of_strings)
def format_missing(missing_from_a, missing_from_b, name_a, name_b):
res = []
if missing_from_a:
res.append("missing from %s:\n%s"
% (name_a, format_list(sorted(missing_from_a))))
if missing_from_b:
res.append("missing from %s:\n%s"
% (name_b, format_list(sorted(missing_from_b))))
return '\n'.join(res)
# Filesystem/OS utilities
class CommandFailed(Failure):
def __init__(self, command, status, output):
Failure.__init__(self, "%s failed (status %s):\n%s" % (
command, status, output))
def run(command, encoding=None, decode=True, cwd=None):
"""Run a command [cmd, arg1, arg2, ...].
Returns the output (stdout + stderr).
Raises CommandFailed in cases of error.
if not encoding:
encoding = locale.getpreferredencoding()
with open(os.devnull, 'rb') as devnull:
pipe = subprocess.Popen(command, stdin=devnull,
stderr=subprocess.PIPE, cwd=cwd)
except OSError as e:
raise Failure("could not run %s: %s" % (command, e))
output, stderr = pipe.communicate()
if decode:
output = output.decode(encoding)
stderr = stderr.decode(encoding, 'replace')
status = pipe.wait()
if status != 0:
raise CommandFailed(command, status, output + stderr)
return output
def cd(directory):
"""Change the current working directory, temporarily.
Use as a context manager: with cd(d): ...
old_dir = os.getcwd()
def mkdtemp(hint=''):
"""Create a temporary directory, then clean it up.
Use as a context manager: with mkdtemp('-purpose'): ...
dirname = tempfile.mkdtemp(prefix='check-manifest-', suffix=hint)
yield dirname
def chmod_plus(path, add_bits=stat.S_IWUSR):
"""Change a file's mode by adding a few bits.
Like chmod +<bits> <path> in a Unix shell.
os.chmod(path, stat.S_IMODE(os.stat(path).st_mode) | add_bits)
except OSError: # pragma: nocover
pass # well, we tried
def rmtree(path):
"""A version of rmtree that can deal with read-only files and directories.
Needed because the stock shutil.rmtree() fails with an access error
when there are read-only files in the directory on Windows, or when the
directory itself is read-only on Unix.
def onerror(func, path, exc_info):
# Did you know what on Python 3.3 on Windows os.remove() and
# os.unlink() are distinct functions?
if func is os.remove or func is os.unlink or func is os.rmdir:
if sys.platform != 'win32':
chmod_plus(os.path.dirname(path), stat.S_IWUSR | stat.S_IXUSR)
shutil.rmtree(path, onerror=onerror)
def copy_files(filelist, destdir):
"""Copy a list of files to destdir, preserving directory structure.
File names should be relative to the current working directory.
for filename in filelist:
destfile = os.path.join(destdir, filename)
# filename should not be absolute, but let's double-check
assert destfile.startswith(destdir + os.path.sep)
destfiledir = os.path.dirname(destfile)
if not os.path.isdir(destfiledir):
if os.path.isdir(filename):
shutil.copy2(filename, destfile)
def get_one_file_in(dirname):
"""Return the pathname of the one file in a directory.
Raises if the directory has no files or more than one file.
files = os.listdir(dirname)
if len(files) > 1:
raise Failure('More than one file exists in %s:\n%s' %
(dirname, '\n'.join(sorted(files))))
elif not files:
raise Failure('No files found in %s' % dirname)
return os.path.join(dirname, files[0])
def unicodify(filename):
"""Make sure filename is Unicode.
Because the tarfile module on Python 2 doesn't return Unicode.
if isinstance(filename, bytes):
return filename.decode(locale.getpreferredencoding())
return filename
def get_archive_file_list(archive_filename):
"""Return the list of files in an archive.
Supports .tar.gz and .zip.
if archive_filename.endswith('.zip'):
with closing(zipfile.ZipFile(archive_filename)) as zf:
return add_directories(zf.namelist())
elif archive_filename.endswith(('.tar.gz', '.tar.bz2', '.tar')):
with closing( as tf:
return add_directories(list(map(unicodify, tf.getnames())))
ext = os.path.splitext(archive_filename)[-1]
raise Failure('Unrecognized archive type: %s' % ext)
def strip_toplevel_name(filelist):
"""Strip toplevel name from a file list.
>>> strip_toplevel_name(['a', 'a/b', 'a/c', 'a/c/d'])
['b', 'c', 'c/d']
>>> strip_toplevel_name(['a', 'a/', 'a/b', 'a/c', 'a/c/d'])
['b', 'c', 'c/d']
>>> strip_toplevel_name(['a/b', 'a/c', 'a/c/d'])
['b', 'c', 'c/d']
if not filelist:
return filelist
prefix = filelist[0]
if '/' in prefix:
prefix = prefix.partition('/')[0] + '/'
names = filelist
prefix = prefix + '/'
names = filelist[1:]
for name in names:
if not name.startswith(prefix):
raise Failure("File doesn't have the common prefix (%s): %s"
% (name, prefix))
return [name[len(prefix):] for name in names if name != prefix]
def add_prefix_to_each(prefix, filelist):
"""Add a prefix to each name in a file list.
>>> filelist = add_prefix_to_each('foo/bar', ['a', 'b', 'c/d'])
>>> [f.replace(os.path.sep, '/') for f in filelist]
['foo/bar/a', 'foo/bar/b', 'foo/bar/c/d']
return [posixpath.join(prefix, name) for name in filelist]
class VCS(object):
def detect(cls, location):
return os.path.isdir(os.path.join(location, cls.metadata_name))
class Git(VCS):
metadata_name = '.git'
# Git for Windows uses UTF-8 instead of the locale encoding.
# Git on POSIX systems uses the locale encoding.
_encoding = 'UTF-8' if sys.platform == 'win32' else None
def detect(cls, location):
# .git can be a file for submodules
return os.path.exists(os.path.join(location, cls.metadata_name))
def get_versioned_files(cls):
"""List all files versioned by git in the current directory."""
files = cls._git_ls_files()
submodules = cls._list_submodules()
for subdir in submodules:
subdir = os.path.relpath(subdir).replace(os.path.sep, '/')
files += add_prefix_to_each(subdir, cls._git_ls_files(subdir))
return add_directories(files)
def _git_ls_files(cls, cwd=None):
output = run(['git', 'ls-files', '-z'], encoding=cls._encoding, cwd=cwd)
return output.split('\0')[:-1]
def _list_submodules(cls):
# This is incredibly expensive on my Jenkins instance (50 seconds for
# each invocation, even when there are no submodules whatsoever).
# Curiously, I cannot reproduce that in Appveyor, or even on the same
# Jenkins machine but when I run the tests manually. Still, 2-hour
# Jenkins runs are bad, so let's avoid running 'git submodule' when
# there's no .gitmodules file.
if not os.path.exists('.gitmodules'):
return []
return run(['git', 'submodule', '--quiet', 'foreach', '--recursive',
'printf "%s/%s\\n" $toplevel $path'], encoding=cls._encoding).splitlines()
class Mercurial(VCS):
metadata_name = '.hg'
def get_versioned_files():
"""List all files under Mercurial control in the current directory."""
output = run(['hg', 'status', '-ncamd', '.'])
return add_directories(output.splitlines())
class Bazaar(VCS):
metadata_name = '.bzr'
def _get_terminal_encoding(self):
# Python 3.6 lets us name the OEM codepage directly, which is lucky
# because it also breaks our old method of OEM codepage detection
# (PEP-528 changed sys.stdout.encoding to UTF-8).
except LookupError:
else: # pragma: nocover
return 'oem'
# Based on bzrlib.osutils.get_terminal_encoding()
encoding = getattr(sys.stdout, 'encoding', None)
if not encoding:
encoding = getattr(sys.stdin, 'encoding', None)
if encoding == 'cp0': # "no codepage"
encoding = None
# NB: bzrlib falls back on bzrlib.osutils.get_user_encoding(),
# which is like locale.getpreferredencoding() on steroids, and
# also includes a fallback from 'ascii' to 'utf-8' when
# sys.platform is 'darwin'. This is probably something we might
# want to do in run(), but I'll wait for somebody to complain
# first, since I don't have a Mac OS X machine and cannot test.
return encoding
def get_versioned_files(cls):
"""List all files versioned in Bazaar in the current directory."""
encoding = cls._get_terminal_encoding()
output = run(['bzr', 'ls', '-VR'], encoding=encoding)
return output.splitlines()
class Subversion(VCS):
metadata_name = '.svn'
def get_versioned_files(cls):
"""List all files under SVN control in the current directory."""
output = run(['svn', 'st', '-vq', '--xml'], decode=False)
tree = ET.XML(output)
return sorted(entry.get('path') for entry in tree.findall('.//entry')
if cls.is_interesting(entry))
def is_interesting(entry):
"""Is this entry interesting?
``entry`` is an XML node representing one entry of the svn status
XML output. It looks like this::
<entry path="unchanged.txt">
<wc-status item="normal" revision="1" props="none">
<commit revision="1">
<entry path="added-but-not-committed.txt">
<wc-status item="added" revision="-1" props="none"></wc-status>
<entry path="ext">
<wc-status item="external" props="none"></wc-status>
<entry path="unknown.txt">
<wc-status props="none" item="unversioned"></wc-status>
if entry.get('path') == '.':
return False
status = entry.find('wc-status')
if status is None:
warning('svn status --xml parse error: <entry path="%s"> without'
' <wc-status>' % entry.get('path'))
return False
# For SVN externals we get two entries: one mentioning the
# existence of the external, and one about the status of the external.
if status.get('item') in ('unversioned', 'external'):
return False
return True
def detect_vcs():
"""Detect the version control system used for the current directory."""
location = os.path.abspath('.')
while True:
for vcs in Git, Mercurial, Bazaar, Subversion:
if vcs.detect(location):
return vcs
parent = os.path.dirname(location)
if parent == location:
raise Failure("Couldn't find version control data"
" (git/hg/bzr/svn supported)")
location = parent
def get_vcs_files():
"""List all files under version control in the current directory."""
vcs = detect_vcs()
return normalize_names(vcs.get_versioned_files())
def normalize_names(names):
"""Normalize file names."""
return list(map(normalize_name, names))
def normalize_name(name):
"""Some VCS print directory names with trailing slashes. Strip them.
Easiest is to normalize the path.
And encodings may trip us up too, especially when comparing lists
of files. Plus maybe lowercase versus uppercase.
name = os.path.normpath(name)
name = unicodify(name)
if sys.platform == 'darwin':
# Mac OSX may have problems comparing non-ascii filenames, so
# we convert them.
name = unicodedata.normalize('NFC', name)
return name
def add_directories(names):
"""Git/Mercurial/zip files omit directories, let's add them back."""
res = list(names)
seen = set(names)
for name in names:
while True:
name = os.path.dirname(name)
if not name or name in seen:
return sorted(res)
# Packaging logic
# it's fine if any of these are missing in the VCS or in the sdist
'PKG-INFO', # always generated
'*.egg-info', # always generated
'*.egg-info/*', # always generated
'setup.cfg', # always generated, sometimes also kept in source control
# it's not a problem if the sdist is lacking these files:
'.hgtags', '.hgsigs', '.hgignore', '.gitignore', '.bzrignore',
'.github', # GitHub template files
'.github/*', # GitHub template files
# it's convenient to ship compiled .mo files in sdists, but they shouldn't
# be checked in
# Regular expressions for filename to ignore. This is useful for
# filename patterns where the '*' part must not search in
# directories.
# generated files should not be committed into the VCS
_sep = r'\\' if os.path.sep == '\\' else os.path.sep
SUGGESTIONS = [(re.compile(pattern.replace('/', _sep)), suggestion) for pattern, suggestion in [
# regexp -> suggestion
('^([^/]+[.](cfg|ini))$', r'include \1'),
('^([.]travis[.]yml)$', r'include \1'),
('^([.]coveragerc)$', r'include \1'),
('^([A-Z]+)$', r'include \1'),
('^(Makefile)$', r'include \1'),
('^[^/]+[.](txt|rst|py)$', r'include *.\1'),
r'recursive-include \1 *.\2'),
r'recursive-include \1 \2'),
# catch-all rules that actually cover some of the above; somewhat
# experimental: I fear false positives
('^([a-zA-Z_0-9]+)$', r'include \1'),
('^[^/]+[.]([a-zA-Z_0-9]+)$', r'include *.\1'),
r'recursive-include \1 *.\2'),
def read_config():
"""Read configuration from file if possible."""
# XXX modifies global state, which is kind of evil
config = _load_config()
if config.get(CFG_IGNORE_DEFAULT_RULES[1], False):
del IGNORE[:]
if CFG_IGNORE[1] in config:
IGNORE.extend(p for p in config[CFG_IGNORE[1]] if p)
if CFG_IGNORE_BAD_IDEAS[1] in config:
IGNORE_BAD_IDEAS.extend(p for p in config[CFG_IGNORE_BAD_IDEAS[1]] if p)
def _load_config():
"""Searches for config files, reads them and returns a dictionary
Looks for a ``check-manifest`` section in ``pyproject.toml``,
``setup.cfg``, and ``tox.ini``, in that order. The first file
that exists and has that section will be loaded and returned as a
if os.path.exists("pyproject.toml"):
config = toml.load("pyproject.toml")
if CFG_SECTION_CHECK_MANIFEST in config.get("tool", {}):
return config["tool"][CFG_SECTION_CHECK_MANIFEST]
search_files = ['setup.cfg', 'tox.ini']
config_parser = ConfigParser.ConfigParser()
for filename in search_files:
if ([filename])
and config_parser.has_section(CFG_SECTION_CHECK_MANIFEST)):
config = {}
if config_parser.has_option(*CFG_IGNORE_DEFAULT_RULES):
ignore_defaults = config_parser.getboolean(*CFG_IGNORE_DEFAULT_RULES)
config[CFG_IGNORE_DEFAULT_RULES[1]] = ignore_defaults
if config_parser.has_option(*CFG_IGNORE):
patterns = [
for p in config_parser.get(*CFG_IGNORE).splitlines()
config[CFG_IGNORE[1]] = patterns
if config_parser.has_option(*CFG_IGNORE_BAD_IDEAS):
patterns = [
for p in config_parser.get(*CFG_IGNORE_BAD_IDEAS).splitlines()
config[CFG_IGNORE_BAD_IDEAS[1]] = patterns
return config
return {}
def read_manifest():
"""Read existing configuration from
We use that to ignore anything the ignores.
# XXX modifies global state, which is kind of evil
if not os.path.isfile(''):
ignore, ignore_regexps = _get_ignore_from_manifest('')
def _glob_to_regexp(pat):
"""Compile a glob pattern into a regexp.
We need to do this because fnmatch allows * to match /, which we
don't want. E.g. a exclude of 'dirname/*css' should
match 'dirname/foo.css' but not 'dirname/subdir/bar.css'.
return glob_to_re(pat)
def _get_ignore_from_manifest(filename):
"""Gather the various ignore patterns from a
Returns a list of standard ignore patterns and a list of regular
expressions to ignore.
class MyTextFile(TextFile):
def error(self, msg, line=None): # pragma: nocover
# (this is never called by TextFile in current versions of CPython)
raise Failure(self.gen_error(msg, line))
def warn(self, msg, line=None):
warning(self.gen_error(msg, line))
template = MyTextFile(filename,
lines = template.readlines()
return _get_ignore_from_manifest_lines(lines)
def _get_ignore_from_manifest_lines(lines):
"""Gather the various ignore patterns from a
'lines' should be a list of strings with comments removed
and continuation lines joined.
Returns a list of standard ignore patterns and a list of regular
expressions to ignore.
ignore = []
ignore_regexps = []
for line in lines:
cmd, rest = line.split(None, 1)
except ValueError:
# no whitespace, so not interesting
for part in rest.split():
# distutils enforces these warnings on Windows only
if part.startswith('/'):
warning("ERROR: Leading slashes are not allowed in on Windows: %s" % part)
if part.endswith('/'):
warning("ERROR: Trailing slashes are not allowed in on Windows: %s" % part)
if cmd == 'exclude':
# An exclude of 'dirname/*css' can match 'dirname/foo.css'
# but not 'dirname/subdir/bar.css'. We need a regular
# expression for that, since fnmatch doesn't pay attention to
# directory separators.
for pat in rest.split():
if '*' in pat or '?' in pat or '[!' in pat:
# No need for special handling.
elif cmd == 'global-exclude':
elif cmd == 'recursive-exclude':
dirname, patterns = rest.split(None, 1)
except ValueError:
# Wrong line.
warning("You have a wrong line in %r\n"
"'recursive-exclude' expects <dir> <pattern1> "
"<pattern2> ..." % line)
# Strip path separator for clarity.
dirname = dirname.rstrip(os.path.sep)
for pattern in patterns.split():
if pattern.startswith('*'):
ignore.append(dirname + os.path.sep + pattern)
# 'recursive-exclude plone metadata.xml' should
# exclude plone/metadata.xml and
# plone/*/metadata.xml, where * can be any number
# of sub directories. We could use a regexp, but
# two ignores seems easier.
ignore.append(dirname + os.path.sep + pattern)
dirname + os.path.sep + '*' + os.path.sep + pattern)
elif cmd == 'prune':
# rest is considered to be a directory name. It should
# not contain a path separator, as it actually has no
# effect in that case, but that could differ per python
# version. We strip it here to avoid double separators.
# XXX: mg: I'm not 100% sure the above is correct, AFAICS
# all pythons from 2.6 complain if the path has a leading or
# trailing slash -- on Windows, that is.
rest = rest.rstrip('/\\')
ignore.append(rest + os.path.sep + '*')
return ignore, ignore_regexps
def file_matches(filename, patterns):
"""Does this filename match any of the patterns?"""
return any(fnmatch.fnmatch(filename, pat)
or fnmatch.fnmatch(os.path.basename(filename), pat)
for pat in patterns)
def file_matches_regexps(filename, patterns):
"""Does this filename match any of the regular expressions?"""
return any(re.match(pat, filename) for pat in patterns)
def strip_sdist_extras(filelist):
"""Strip generated files that are only present in source distributions.
We also strip files that are ignored for other reasons, like
command line arguments, setup.cfg rules or rules.
return [name for name in filelist
if not file_matches(name, IGNORE)
and not file_matches_regexps(name, IGNORE_REGEXPS)]
def find_bad_ideas(filelist):
"""Find files matching WARN_ABOUT_FILES_IN_VCS patterns."""
return [name for name in filelist
if file_matches(name, WARN_ABOUT_FILES_IN_VCS)]
def find_suggestions(filelist):
"""Suggest patterns for missing files."""
suggestions = set()
unknowns = []
for filename in filelist:
if os.path.isdir(filename):
# it's impossible to add empty directories via anyway,
# and non-empty directories will be added automatically when we
# specify patterns for files inside them
for pattern, suggestion in SUGGESTIONS:
m = pattern.match(filename)
if m is not None:
suggestions.add(pattern.sub(suggestion, filename))
return sorted(suggestions), unknowns
def is_package(source_tree='.'):
"""Is the directory the root of a Python package?
Note: the term "package" here refers to a collection of files
with a, not to a directory with an
return (
os.path.exists(os.path.join(source_tree, ''))
or os.path.exists(os.path.join(source_tree, 'pyproject.toml'))
def extract_version_from_filename(filename):
"""Extract version number from sdist filename."""
filename = os.path.splitext(os.path.basename(filename))[0]
if filename.endswith('.tar'):
filename = os.path.splitext(filename)[0]
return filename.partition('-')[2]
def should_use_pep_517():
"""Check if the project uses PEP-517 builds."""
# says
# "If the pyproject.toml file is absent, or the build-backend key is
# missing, the source tree is not using this specification, and tools
# should revert to the legacy behaviour of running".
if not os.path.exists('pyproject.toml'):
return False
config = toml.load("pyproject.toml")
if "build-system" not in config:
return False
if "build-backend" not in config["build-system"]:
return False
return True
def build_sdist(tempdir, python=sys.executable):
"""Build a source distribution in a temporary directory.
Should be run with the current working directory inside the Python package
you want to build.
if should_use_pep_517():
# I could do this in-process with
# import pep517.envbuild
# pep517.envbuild.build_sdist('.', tempdir)
# but then it would print a bunch of things to stdout and I'd have to
# worry about exceptions
run([python, '-m', '', '--source', '-o', tempdir, '.'])
run([python, '', 'sdist', '-d', tempdir])
def check_manifest(source_tree='.', create=False, update=False,
"""Compare a generated source distribution with list of files in a VCS.
Returns True if the manifest is fine.
all_ok = True
if os.path.sep in python:
python = os.path.abspath(python)
with cd(source_tree):
if not is_package():
raise Failure(
'This is not a Python project (no')
info_begin("listing source files under version control")
all_source_files = sorted(get_vcs_files())
source_files = strip_sdist_extras(all_source_files)
info_continue(": %d files and directories" % len(source_files))
if not all_source_files:
raise Failure('There are no files added to version control!')
info_begin("building an sdist")
with mkdtemp('-sdist') as tempdir:
build_sdist(tempdir, python=python)
sdist_filename = get_one_file_in(tempdir)
info_continue(": %s" % os.path.basename(sdist_filename))
sdist_files = sorted(normalize_names(strip_sdist_extras(
info_continue(": %d files and directories" % len(sdist_files))
version = extract_version_from_filename(sdist_filename)
existing_source_files = list(filter(os.path.exists, all_source_files))
missing_source_files = sorted(set(all_source_files) - set(existing_source_files))
if missing_source_files:
warning("some files listed as being under source control are missing:\n%s"
% format_list(missing_source_files))
info_begin("copying source files to a temporary directory")
with mkdtemp('-sources') as tempsourcedir:
copy_files(existing_source_files, tempsourcedir)
for filename in '', '', 'pyproject.toml':
if filename not in source_files and os.path.exists(filename):
# See
# and
# if we do this, the user gets a warning about files
# missing from source control; if we don't do this,
# things get very confusing for the user!
copy_files([filename], tempsourcedir)
info_begin("building a clean sdist")
with cd(tempsourcedir):
with mkdtemp('-sdist') as tempdir:
os.environ['SETUPTOOLS_SCM_PRETEND_VERSION'] = version
build_sdist(tempdir, python=python)
sdist_filename = get_one_file_in(tempdir)
info_continue(": %s" % os.path.basename(sdist_filename))
clean_sdist_files = sorted(normalize_names(strip_sdist_extras(
info_continue(": %d files and directories" % len(clean_sdist_files))
missing_from_manifest = set(source_files) - set(clean_sdist_files)
missing_from_VCS = set(sdist_files + clean_sdist_files) - set(source_files)
if not missing_from_manifest and not missing_from_VCS:
info("lists of files in version control and sdist match")
error("lists of files in version control and sdist do not match!\n%s"
% format_missing(missing_from_VCS, missing_from_manifest,
"VCS", "sdist"))
suggestions, unknowns = find_suggestions(missing_from_manifest)
user_asked_for_help = update or (create and not
if '' not in existing_source_files:
if suggestions and not user_asked_for_help:
info("no found; you can run 'check-manifest -c' to create one")
info("no found")
if suggestions:
info("suggested rules:\n%s"
% format_list(suggestions))
if user_asked_for_help:
existed = os.path.exists('')
with open('', 'a') as f:
if not existed:
f.write('\n# added by\n')
f.write('\n'.join(suggestions) + '\n')
if unknowns:
info("don't know how to come up with rules matching\n%s"
% format_list(unknowns))
elif user_asked_for_help:
info("don't know how to come up with rules"
" matching any of the files, sorry!")
all_ok = False
bad_ideas = find_bad_ideas(all_source_files)
filtered_bad_ideas = [bad_idea for bad_idea in bad_ideas
if not file_matches(bad_idea, IGNORE_BAD_IDEAS)]
if filtered_bad_ideas:
warning("you have %s in source control!\nthat's a bad idea:"
" auto-generated files should not be versioned"
% filtered_bad_ideas[0])
if len(filtered_bad_ideas) > 1:
warning("this also applies to the following:\n%s"
% format_list(filtered_bad_ideas[1:]))
all_ok = False
return all_ok
# Main script
def main():
parser = argparse.ArgumentParser(
description="Check a Python file for completeness",
parser.add_argument('source_tree', default='.', nargs='?',
help='location for the source tree')
parser.add_argument('--version', action='version',
version='%(prog)s version ' + __version__)
parser.add_argument('-v', '--verbose', action='store_true',
help='more verbose output')
parser.add_argument('-c', '--create', action='store_true',
help='create a if missing')
parser.add_argument('-u', '--update', action='store_true',
help='append suggestions to (implies --create)')
parser.add_argument('-p', '--python', default=sys.executable,
help='use this Python interpreter for running sdist')
parser.add_argument('--ignore', metavar='patterns', default=None,
help='ignore files/directories matching these'
' comma-separated patterns')
parser.add_argument('--ignore-bad-ideas', metavar='patterns',
default=[], help='ignore bad idea files/directories '
'matching these comma-separated patterns')
args = parser.parse_args()
if args.ignore:
if args.ignore_bad_ideas:
if args.verbose:
global VERBOSE
if not check_manifest(args.source_tree, create=args.create,
update=args.update, python=args.python):
except Failure as e:
# zest.releaser integration
def zest_releaser_check(data):
"""Check the completeness of before the release.
This is an entry point for zest.releaser. See the documentation at
from zest.releaser.utils import ask
source_tree = data['workingdir']
if not is_package(source_tree):
# You can use zest.releaser on things that are not Python packages.
# It's pointless to run check-manifest in those circumstances.
# See for details.
if not ask("Do you want to run check-manifest?"):
if not check_manifest(source_tree):
if not ask(" has problems. "
" Do you want to continue despite that?", default=False):
except Failure as e:
if not ask("Something bad happened. "
" Do you want to continue despite that?", default=False):
if __name__ == '__main__':