1235 lines
45 KiB
Python
1235 lines
45 KiB
Python
# -*- coding: utf-8 -*-
|
|
__all__ = ['Distribution']
|
|
|
|
import io
|
|
import sys
|
|
import re
|
|
import os
|
|
import warnings
|
|
import numbers
|
|
import distutils.log
|
|
import distutils.core
|
|
import distutils.cmd
|
|
import distutils.dist
|
|
import distutils.command
|
|
from distutils.util import strtobool
|
|
from distutils.debug import DEBUG
|
|
from distutils.fancy_getopt import translate_longopt
|
|
from glob import iglob
|
|
import itertools
|
|
import textwrap
|
|
from typing import List, Optional, TYPE_CHECKING
|
|
from pathlib import Path
|
|
|
|
from collections import defaultdict
|
|
from email import message_from_file
|
|
|
|
from distutils.errors import DistutilsOptionError, DistutilsSetupError
|
|
from distutils.util import rfc822_escape
|
|
|
|
from setuptools.extern import packaging
|
|
from setuptools.extern import ordered_set
|
|
from setuptools.extern.more_itertools import unique_everseen, partition
|
|
from setuptools.extern import nspektr
|
|
|
|
from ._importlib import metadata
|
|
|
|
from . import SetuptoolsDeprecationWarning
|
|
|
|
import setuptools
|
|
import setuptools.command
|
|
from setuptools import windows_support
|
|
from setuptools.monkey import get_unpatched
|
|
from setuptools.config import setupcfg, pyprojecttoml
|
|
from setuptools.discovery import ConfigDiscovery
|
|
|
|
import pkg_resources
|
|
from setuptools.extern.packaging import version
|
|
from . import _reqs
|
|
from . import _entry_points
|
|
|
|
if TYPE_CHECKING:
|
|
from email.message import Message
|
|
|
|
__import__('setuptools.extern.packaging.specifiers')
|
|
__import__('setuptools.extern.packaging.version')
|
|
|
|
|
|
def _get_unpatched(cls):
|
|
warnings.warn("Do not call this function", DistDeprecationWarning)
|
|
return get_unpatched(cls)
|
|
|
|
|
|
def get_metadata_version(self):
|
|
mv = getattr(self, 'metadata_version', None)
|
|
if mv is None:
|
|
mv = version.Version('2.1')
|
|
self.metadata_version = mv
|
|
return mv
|
|
|
|
|
|
def rfc822_unescape(content: str) -> str:
|
|
"""Reverse RFC-822 escaping by removing leading whitespaces from content."""
|
|
lines = content.splitlines()
|
|
if len(lines) == 1:
|
|
return lines[0].lstrip()
|
|
return '\n'.join((lines[0].lstrip(), textwrap.dedent('\n'.join(lines[1:]))))
|
|
|
|
|
|
def _read_field_from_msg(msg: "Message", field: str) -> Optional[str]:
|
|
"""Read Message header field."""
|
|
value = msg[field]
|
|
if value == 'UNKNOWN':
|
|
return None
|
|
return value
|
|
|
|
|
|
def _read_field_unescaped_from_msg(msg: "Message", field: str) -> Optional[str]:
|
|
"""Read Message header field and apply rfc822_unescape."""
|
|
value = _read_field_from_msg(msg, field)
|
|
if value is None:
|
|
return value
|
|
return rfc822_unescape(value)
|
|
|
|
|
|
def _read_list_from_msg(msg: "Message", field: str) -> Optional[List[str]]:
|
|
"""Read Message header field and return all results as list."""
|
|
values = msg.get_all(field, None)
|
|
if values == []:
|
|
return None
|
|
return values
|
|
|
|
|
|
def _read_payload_from_msg(msg: "Message") -> Optional[str]:
|
|
value = msg.get_payload().strip()
|
|
if value == 'UNKNOWN' or not value:
|
|
return None
|
|
return value
|
|
|
|
|
|
def read_pkg_file(self, file):
|
|
"""Reads the metadata values from a file object."""
|
|
msg = message_from_file(file)
|
|
|
|
self.metadata_version = version.Version(msg['metadata-version'])
|
|
self.name = _read_field_from_msg(msg, 'name')
|
|
self.version = _read_field_from_msg(msg, 'version')
|
|
self.description = _read_field_from_msg(msg, 'summary')
|
|
# we are filling author only.
|
|
self.author = _read_field_from_msg(msg, 'author')
|
|
self.maintainer = None
|
|
self.author_email = _read_field_from_msg(msg, 'author-email')
|
|
self.maintainer_email = None
|
|
self.url = _read_field_from_msg(msg, 'home-page')
|
|
self.download_url = _read_field_from_msg(msg, 'download-url')
|
|
self.license = _read_field_unescaped_from_msg(msg, 'license')
|
|
|
|
self.long_description = _read_field_unescaped_from_msg(msg, 'description')
|
|
if (
|
|
self.long_description is None and
|
|
self.metadata_version >= version.Version('2.1')
|
|
):
|
|
self.long_description = _read_payload_from_msg(msg)
|
|
self.description = _read_field_from_msg(msg, 'summary')
|
|
|
|
if 'keywords' in msg:
|
|
self.keywords = _read_field_from_msg(msg, 'keywords').split(',')
|
|
|
|
self.platforms = _read_list_from_msg(msg, 'platform')
|
|
self.classifiers = _read_list_from_msg(msg, 'classifier')
|
|
|
|
# PEP 314 - these fields only exist in 1.1
|
|
if self.metadata_version == version.Version('1.1'):
|
|
self.requires = _read_list_from_msg(msg, 'requires')
|
|
self.provides = _read_list_from_msg(msg, 'provides')
|
|
self.obsoletes = _read_list_from_msg(msg, 'obsoletes')
|
|
else:
|
|
self.requires = None
|
|
self.provides = None
|
|
self.obsoletes = None
|
|
|
|
self.license_files = _read_list_from_msg(msg, 'license-file')
|
|
|
|
|
|
def single_line(val):
|
|
"""
|
|
Quick and dirty validation for Summary pypa/setuptools#1390.
|
|
"""
|
|
if '\n' in val:
|
|
# TODO: Replace with `raise ValueError("newlines not allowed")`
|
|
# after reviewing #2893.
|
|
warnings.warn("newlines not allowed and will break in the future")
|
|
val = val.strip().split('\n')[0]
|
|
return val
|
|
|
|
|
|
# Based on Python 3.5 version
|
|
def write_pkg_file(self, file): # noqa: C901 # is too complex (14) # FIXME
|
|
"""Write the PKG-INFO format data to a file object."""
|
|
version = self.get_metadata_version()
|
|
|
|
def write_field(key, value):
|
|
file.write("%s: %s\n" % (key, value))
|
|
|
|
write_field('Metadata-Version', str(version))
|
|
write_field('Name', self.get_name())
|
|
write_field('Version', self.get_version())
|
|
|
|
summary = self.get_description()
|
|
if summary:
|
|
write_field('Summary', single_line(summary))
|
|
|
|
optional_fields = (
|
|
('Home-page', 'url'),
|
|
('Download-URL', 'download_url'),
|
|
('Author', 'author'),
|
|
('Author-email', 'author_email'),
|
|
('Maintainer', 'maintainer'),
|
|
('Maintainer-email', 'maintainer_email'),
|
|
)
|
|
|
|
for field, attr in optional_fields:
|
|
attr_val = getattr(self, attr, None)
|
|
if attr_val is not None:
|
|
write_field(field, attr_val)
|
|
|
|
license = self.get_license()
|
|
if license:
|
|
write_field('License', rfc822_escape(license))
|
|
|
|
for project_url in self.project_urls.items():
|
|
write_field('Project-URL', '%s, %s' % project_url)
|
|
|
|
keywords = ','.join(self.get_keywords())
|
|
if keywords:
|
|
write_field('Keywords', keywords)
|
|
|
|
platforms = self.get_platforms() or []
|
|
for platform in platforms:
|
|
write_field('Platform', platform)
|
|
|
|
self._write_list(file, 'Classifier', self.get_classifiers())
|
|
|
|
# PEP 314
|
|
self._write_list(file, 'Requires', self.get_requires())
|
|
self._write_list(file, 'Provides', self.get_provides())
|
|
self._write_list(file, 'Obsoletes', self.get_obsoletes())
|
|
|
|
# Setuptools specific for PEP 345
|
|
if hasattr(self, 'python_requires'):
|
|
write_field('Requires-Python', self.python_requires)
|
|
|
|
# PEP 566
|
|
if self.long_description_content_type:
|
|
write_field('Description-Content-Type', self.long_description_content_type)
|
|
if self.provides_extras:
|
|
for extra in self.provides_extras:
|
|
write_field('Provides-Extra', extra)
|
|
|
|
self._write_list(file, 'License-File', self.license_files or [])
|
|
|
|
long_description = self.get_long_description()
|
|
if long_description:
|
|
file.write("\n%s" % long_description)
|
|
if not long_description.endswith("\n"):
|
|
file.write("\n")
|
|
|
|
|
|
sequence = tuple, list
|
|
|
|
|
|
def check_importable(dist, attr, value):
|
|
try:
|
|
ep = metadata.EntryPoint(value=value, name=None, group=None)
|
|
assert not ep.extras
|
|
except (TypeError, ValueError, AttributeError, AssertionError) as e:
|
|
raise DistutilsSetupError(
|
|
"%r must be importable 'module:attrs' string (got %r)" % (attr, value)
|
|
) from e
|
|
|
|
|
|
def assert_string_list(dist, attr, value):
|
|
"""Verify that value is a string list"""
|
|
try:
|
|
# verify that value is a list or tuple to exclude unordered
|
|
# or single-use iterables
|
|
assert isinstance(value, (list, tuple))
|
|
# verify that elements of value are strings
|
|
assert ''.join(value) != value
|
|
except (TypeError, ValueError, AttributeError, AssertionError) as e:
|
|
raise DistutilsSetupError(
|
|
"%r must be a list of strings (got %r)" % (attr, value)
|
|
) from e
|
|
|
|
|
|
def check_nsp(dist, attr, value):
|
|
"""Verify that namespace packages are valid"""
|
|
ns_packages = value
|
|
assert_string_list(dist, attr, ns_packages)
|
|
for nsp in ns_packages:
|
|
if not dist.has_contents_for(nsp):
|
|
raise DistutilsSetupError(
|
|
"Distribution contains no modules or packages for "
|
|
+ "namespace package %r" % nsp
|
|
)
|
|
parent, sep, child = nsp.rpartition('.')
|
|
if parent and parent not in ns_packages:
|
|
distutils.log.warn(
|
|
"WARNING: %r is declared as a package namespace, but %r"
|
|
" is not: please correct this in setup.py",
|
|
nsp,
|
|
parent,
|
|
)
|
|
msg = (
|
|
"The namespace_packages parameter is deprecated, "
|
|
"consider using implicit namespaces instead (PEP 420)."
|
|
)
|
|
warnings.warn(msg, SetuptoolsDeprecationWarning)
|
|
|
|
|
|
def check_extras(dist, attr, value):
|
|
"""Verify that extras_require mapping is valid"""
|
|
try:
|
|
list(itertools.starmap(_check_extra, value.items()))
|
|
except (TypeError, ValueError, AttributeError) as e:
|
|
raise DistutilsSetupError(
|
|
"'extras_require' must be a dictionary whose values are "
|
|
"strings or lists of strings containing valid project/version "
|
|
"requirement specifiers."
|
|
) from e
|
|
|
|
|
|
def _check_extra(extra, reqs):
|
|
name, sep, marker = extra.partition(':')
|
|
if marker and pkg_resources.invalid_marker(marker):
|
|
raise DistutilsSetupError("Invalid environment marker: " + marker)
|
|
list(_reqs.parse(reqs))
|
|
|
|
|
|
def assert_bool(dist, attr, value):
|
|
"""Verify that value is True, False, 0, or 1"""
|
|
if bool(value) != value:
|
|
tmpl = "{attr!r} must be a boolean value (got {value!r})"
|
|
raise DistutilsSetupError(tmpl.format(attr=attr, value=value))
|
|
|
|
|
|
def invalid_unless_false(dist, attr, value):
|
|
if not value:
|
|
warnings.warn(f"{attr} is ignored.", DistDeprecationWarning)
|
|
return
|
|
raise DistutilsSetupError(f"{attr} is invalid.")
|
|
|
|
|
|
def check_requirements(dist, attr, value):
|
|
"""Verify that install_requires is a valid requirements list"""
|
|
try:
|
|
list(_reqs.parse(value))
|
|
if isinstance(value, (dict, set)):
|
|
raise TypeError("Unordered types are not allowed")
|
|
except (TypeError, ValueError) as error:
|
|
tmpl = (
|
|
"{attr!r} must be a string or list of strings "
|
|
"containing valid project/version requirement specifiers; {error}"
|
|
)
|
|
raise DistutilsSetupError(tmpl.format(attr=attr, error=error)) from error
|
|
|
|
|
|
def check_specifier(dist, attr, value):
|
|
"""Verify that value is a valid version specifier"""
|
|
try:
|
|
packaging.specifiers.SpecifierSet(value)
|
|
except (packaging.specifiers.InvalidSpecifier, AttributeError) as error:
|
|
tmpl = (
|
|
"{attr!r} must be a string " "containing valid version specifiers; {error}"
|
|
)
|
|
raise DistutilsSetupError(tmpl.format(attr=attr, error=error)) from error
|
|
|
|
|
|
def check_entry_points(dist, attr, value):
|
|
"""Verify that entry_points map is parseable"""
|
|
try:
|
|
_entry_points.load(value)
|
|
except Exception as e:
|
|
raise DistutilsSetupError(e) from e
|
|
|
|
|
|
def check_test_suite(dist, attr, value):
|
|
if not isinstance(value, str):
|
|
raise DistutilsSetupError("test_suite must be a string")
|
|
|
|
|
|
def check_package_data(dist, attr, value):
|
|
"""Verify that value is a dictionary of package names to glob lists"""
|
|
if not isinstance(value, dict):
|
|
raise DistutilsSetupError(
|
|
"{!r} must be a dictionary mapping package names to lists of "
|
|
"string wildcard patterns".format(attr)
|
|
)
|
|
for k, v in value.items():
|
|
if not isinstance(k, str):
|
|
raise DistutilsSetupError(
|
|
"keys of {!r} dict must be strings (got {!r})".format(attr, k)
|
|
)
|
|
assert_string_list(dist, 'values of {!r} dict'.format(attr), v)
|
|
|
|
|
|
def check_packages(dist, attr, value):
|
|
for pkgname in value:
|
|
if not re.match(r'\w+(\.\w+)*', pkgname):
|
|
distutils.log.warn(
|
|
"WARNING: %r not a valid package name; please use only "
|
|
".-separated package names in setup.py",
|
|
pkgname,
|
|
)
|
|
|
|
|
|
_Distribution = get_unpatched(distutils.core.Distribution)
|
|
|
|
|
|
class Distribution(_Distribution):
|
|
"""Distribution with support for tests and package data
|
|
|
|
This is an enhanced version of 'distutils.dist.Distribution' that
|
|
effectively adds the following new optional keyword arguments to 'setup()':
|
|
|
|
'install_requires' -- a string or sequence of strings specifying project
|
|
versions that the distribution requires when installed, in the format
|
|
used by 'pkg_resources.require()'. They will be installed
|
|
automatically when the package is installed. If you wish to use
|
|
packages that are not available in PyPI, or want to give your users an
|
|
alternate download location, you can add a 'find_links' option to the
|
|
'[easy_install]' section of your project's 'setup.cfg' file, and then
|
|
setuptools will scan the listed web pages for links that satisfy the
|
|
requirements.
|
|
|
|
'extras_require' -- a dictionary mapping names of optional "extras" to the
|
|
additional requirement(s) that using those extras incurs. For example,
|
|
this::
|
|
|
|
extras_require = dict(reST = ["docutils>=0.3", "reSTedit"])
|
|
|
|
indicates that the distribution can optionally provide an extra
|
|
capability called "reST", but it can only be used if docutils and
|
|
reSTedit are installed. If the user installs your package using
|
|
EasyInstall and requests one of your extras, the corresponding
|
|
additional requirements will be installed if needed.
|
|
|
|
'test_suite' -- the name of a test suite to run for the 'test' command.
|
|
If the user runs 'python setup.py test', the package will be installed,
|
|
and the named test suite will be run. The format is the same as
|
|
would be used on a 'unittest.py' command line. That is, it is the
|
|
dotted name of an object to import and call to generate a test suite.
|
|
|
|
'package_data' -- a dictionary mapping package names to lists of filenames
|
|
or globs to use to find data files contained in the named packages.
|
|
If the dictionary has filenames or globs listed under '""' (the empty
|
|
string), those names will be searched for in every package, in addition
|
|
to any names for the specific package. Data files found using these
|
|
names/globs will be installed along with the package, in the same
|
|
location as the package. Note that globs are allowed to reference
|
|
the contents of non-package subdirectories, as long as you use '/' as
|
|
a path separator. (Globs are automatically converted to
|
|
platform-specific paths at runtime.)
|
|
|
|
In addition to these new keywords, this class also has several new methods
|
|
for manipulating the distribution's contents. For example, the 'include()'
|
|
and 'exclude()' methods can be thought of as in-place add and subtract
|
|
commands that add or remove packages, modules, extensions, and so on from
|
|
the distribution.
|
|
"""
|
|
|
|
_DISTUTILS_UNSUPPORTED_METADATA = {
|
|
'long_description_content_type': lambda: None,
|
|
'project_urls': dict,
|
|
'provides_extras': ordered_set.OrderedSet,
|
|
'license_file': lambda: None,
|
|
'license_files': lambda: None,
|
|
}
|
|
|
|
_patched_dist = None
|
|
|
|
def patch_missing_pkg_info(self, attrs):
|
|
# Fake up a replacement for the data that would normally come from
|
|
# PKG-INFO, but which might not yet be built if this is a fresh
|
|
# checkout.
|
|
#
|
|
if not attrs or 'name' not in attrs or 'version' not in attrs:
|
|
return
|
|
key = pkg_resources.safe_name(str(attrs['name'])).lower()
|
|
dist = pkg_resources.working_set.by_key.get(key)
|
|
if dist is not None and not dist.has_metadata('PKG-INFO'):
|
|
dist._version = pkg_resources.safe_version(str(attrs['version']))
|
|
self._patched_dist = dist
|
|
|
|
def __init__(self, attrs=None):
|
|
have_package_data = hasattr(self, "package_data")
|
|
if not have_package_data:
|
|
self.package_data = {}
|
|
attrs = attrs or {}
|
|
self.dist_files = []
|
|
# Filter-out setuptools' specific options.
|
|
self.src_root = attrs.pop("src_root", None)
|
|
self.patch_missing_pkg_info(attrs)
|
|
self.dependency_links = attrs.pop('dependency_links', [])
|
|
self.setup_requires = attrs.pop('setup_requires', [])
|
|
for ep in metadata.entry_points(group='distutils.setup_keywords'):
|
|
vars(self).setdefault(ep.name, None)
|
|
_Distribution.__init__(
|
|
self,
|
|
{
|
|
k: v
|
|
for k, v in attrs.items()
|
|
if k not in self._DISTUTILS_UNSUPPORTED_METADATA
|
|
},
|
|
)
|
|
|
|
# Save the original dependencies before they are processed into the egg format
|
|
self._orig_extras_require = {}
|
|
self._orig_install_requires = []
|
|
self._tmp_extras_require = defaultdict(ordered_set.OrderedSet)
|
|
|
|
self.set_defaults = ConfigDiscovery(self)
|
|
|
|
self._set_metadata_defaults(attrs)
|
|
|
|
self.metadata.version = self._normalize_version(
|
|
self._validate_version(self.metadata.version)
|
|
)
|
|
self._finalize_requires()
|
|
|
|
def _validate_metadata(self):
|
|
required = {"name"}
|
|
provided = {
|
|
key
|
|
for key in vars(self.metadata)
|
|
if getattr(self.metadata, key, None) is not None
|
|
}
|
|
missing = required - provided
|
|
|
|
if missing:
|
|
msg = f"Required package metadata is missing: {missing}"
|
|
raise DistutilsSetupError(msg)
|
|
|
|
def _set_metadata_defaults(self, attrs):
|
|
"""
|
|
Fill-in missing metadata fields not supported by distutils.
|
|
Some fields may have been set by other tools (e.g. pbr).
|
|
Those fields (vars(self.metadata)) take precedence to
|
|
supplied attrs.
|
|
"""
|
|
for option, default in self._DISTUTILS_UNSUPPORTED_METADATA.items():
|
|
vars(self.metadata).setdefault(option, attrs.get(option, default()))
|
|
|
|
@staticmethod
|
|
def _normalize_version(version):
|
|
if isinstance(version, setuptools.sic) or version is None:
|
|
return version
|
|
|
|
normalized = str(packaging.version.Version(version))
|
|
if version != normalized:
|
|
tmpl = "Normalizing '{version}' to '{normalized}'"
|
|
warnings.warn(tmpl.format(**locals()))
|
|
return normalized
|
|
return version
|
|
|
|
@staticmethod
|
|
def _validate_version(version):
|
|
if isinstance(version, numbers.Number):
|
|
# Some people apparently take "version number" too literally :)
|
|
version = str(version)
|
|
|
|
if version is not None:
|
|
try:
|
|
packaging.version.Version(version)
|
|
except (packaging.version.InvalidVersion, TypeError):
|
|
warnings.warn(
|
|
"The version specified (%r) is an invalid version, this "
|
|
"may not work as expected with newer versions of "
|
|
"setuptools, pip, and PyPI. Please see PEP 440 for more "
|
|
"details." % version
|
|
)
|
|
return setuptools.sic(version)
|
|
return version
|
|
|
|
def _finalize_requires(self):
|
|
"""
|
|
Set `metadata.python_requires` and fix environment markers
|
|
in `install_requires` and `extras_require`.
|
|
"""
|
|
if getattr(self, 'python_requires', None):
|
|
self.metadata.python_requires = self.python_requires
|
|
|
|
if getattr(self, 'extras_require', None):
|
|
# Save original before it is messed by _convert_extras_requirements
|
|
self._orig_extras_require = self._orig_extras_require or self.extras_require
|
|
for extra in self.extras_require.keys():
|
|
# Since this gets called multiple times at points where the
|
|
# keys have become 'converted' extras, ensure that we are only
|
|
# truly adding extras we haven't seen before here.
|
|
extra = extra.split(':')[0]
|
|
if extra:
|
|
self.metadata.provides_extras.add(extra)
|
|
|
|
if getattr(self, 'install_requires', None) and not self._orig_install_requires:
|
|
# Save original before it is messed by _move_install_requirements_markers
|
|
self._orig_install_requires = self.install_requires
|
|
|
|
self._convert_extras_requirements()
|
|
self._move_install_requirements_markers()
|
|
|
|
def _convert_extras_requirements(self):
|
|
"""
|
|
Convert requirements in `extras_require` of the form
|
|
`"extra": ["barbazquux; {marker}"]` to
|
|
`"extra:{marker}": ["barbazquux"]`.
|
|
"""
|
|
spec_ext_reqs = getattr(self, 'extras_require', None) or {}
|
|
tmp = defaultdict(ordered_set.OrderedSet)
|
|
self._tmp_extras_require = getattr(self, '_tmp_extras_require', tmp)
|
|
for section, v in spec_ext_reqs.items():
|
|
# Do not strip empty sections.
|
|
self._tmp_extras_require[section]
|
|
for r in _reqs.parse(v):
|
|
suffix = self._suffix_for(r)
|
|
self._tmp_extras_require[section + suffix].append(r)
|
|
|
|
@staticmethod
|
|
def _suffix_for(req):
|
|
"""
|
|
For a requirement, return the 'extras_require' suffix for
|
|
that requirement.
|
|
"""
|
|
return ':' + str(req.marker) if req.marker else ''
|
|
|
|
def _move_install_requirements_markers(self):
|
|
"""
|
|
Move requirements in `install_requires` that are using environment
|
|
markers `extras_require`.
|
|
"""
|
|
|
|
# divide the install_requires into two sets, simple ones still
|
|
# handled by install_requires and more complex ones handled
|
|
# by extras_require.
|
|
|
|
def is_simple_req(req):
|
|
return not req.marker
|
|
|
|
spec_inst_reqs = getattr(self, 'install_requires', None) or ()
|
|
inst_reqs = list(_reqs.parse(spec_inst_reqs))
|
|
simple_reqs = filter(is_simple_req, inst_reqs)
|
|
complex_reqs = itertools.filterfalse(is_simple_req, inst_reqs)
|
|
self.install_requires = list(map(str, simple_reqs))
|
|
|
|
for r in complex_reqs:
|
|
self._tmp_extras_require[':' + str(r.marker)].append(r)
|
|
self.extras_require = dict(
|
|
# list(dict.fromkeys(...)) ensures a list of unique strings
|
|
(k, list(dict.fromkeys(str(r) for r in map(self._clean_req, v))))
|
|
for k, v in self._tmp_extras_require.items()
|
|
)
|
|
|
|
def _clean_req(self, req):
|
|
"""
|
|
Given a Requirement, remove environment markers and return it.
|
|
"""
|
|
req.marker = None
|
|
return req
|
|
|
|
def _finalize_license_files(self):
|
|
"""Compute names of all license files which should be included."""
|
|
license_files: Optional[List[str]] = self.metadata.license_files
|
|
patterns: List[str] = license_files if license_files else []
|
|
|
|
license_file: Optional[str] = self.metadata.license_file
|
|
if license_file and license_file not in patterns:
|
|
patterns.append(license_file)
|
|
|
|
if license_files is None and license_file is None:
|
|
# Default patterns match the ones wheel uses
|
|
# See https://wheel.readthedocs.io/en/stable/user_guide.html
|
|
# -> 'Including license files in the generated wheel file'
|
|
patterns = ('LICEN[CS]E*', 'COPYING*', 'NOTICE*', 'AUTHORS*')
|
|
|
|
self.metadata.license_files = list(
|
|
unique_everseen(self._expand_patterns(patterns))
|
|
)
|
|
|
|
@staticmethod
|
|
def _expand_patterns(patterns):
|
|
"""
|
|
>>> list(Distribution._expand_patterns(['LICENSE']))
|
|
['LICENSE']
|
|
>>> list(Distribution._expand_patterns(['setup.cfg', 'LIC*']))
|
|
['setup.cfg', 'LICENSE']
|
|
"""
|
|
return (
|
|
path
|
|
for pattern in patterns
|
|
for path in sorted(iglob(pattern))
|
|
if not path.endswith('~') and os.path.isfile(path)
|
|
)
|
|
|
|
# FIXME: 'Distribution._parse_config_files' is too complex (14)
|
|
def _parse_config_files(self, filenames=None): # noqa: C901
|
|
"""
|
|
Adapted from distutils.dist.Distribution.parse_config_files,
|
|
this method provides the same functionality in subtly-improved
|
|
ways.
|
|
"""
|
|
from configparser import ConfigParser
|
|
|
|
# Ignore install directory options if we have a venv
|
|
ignore_options = (
|
|
[]
|
|
if sys.prefix == sys.base_prefix
|
|
else [
|
|
'install-base',
|
|
'install-platbase',
|
|
'install-lib',
|
|
'install-platlib',
|
|
'install-purelib',
|
|
'install-headers',
|
|
'install-scripts',
|
|
'install-data',
|
|
'prefix',
|
|
'exec-prefix',
|
|
'home',
|
|
'user',
|
|
'root',
|
|
]
|
|
)
|
|
|
|
ignore_options = frozenset(ignore_options)
|
|
|
|
if filenames is None:
|
|
filenames = self.find_config_files()
|
|
|
|
if DEBUG:
|
|
self.announce("Distribution.parse_config_files():")
|
|
|
|
parser = ConfigParser()
|
|
parser.optionxform = str
|
|
for filename in filenames:
|
|
with io.open(filename, encoding='utf-8') as reader:
|
|
if DEBUG:
|
|
self.announce(" reading {filename}".format(**locals()))
|
|
parser.read_file(reader)
|
|
for section in parser.sections():
|
|
options = parser.options(section)
|
|
opt_dict = self.get_option_dict(section)
|
|
|
|
for opt in options:
|
|
if opt == '__name__' or opt in ignore_options:
|
|
continue
|
|
|
|
val = parser.get(section, opt)
|
|
opt = self.warn_dash_deprecation(opt, section)
|
|
opt = self.make_option_lowercase(opt, section)
|
|
opt_dict[opt] = (filename, val)
|
|
|
|
# Make the ConfigParser forget everything (so we retain
|
|
# the original filenames that options come from)
|
|
parser.__init__()
|
|
|
|
if 'global' not in self.command_options:
|
|
return
|
|
|
|
# If there was a "global" section in the config file, use it
|
|
# to set Distribution options.
|
|
|
|
for (opt, (src, val)) in self.command_options['global'].items():
|
|
alias = self.negative_opt.get(opt)
|
|
if alias:
|
|
val = not strtobool(val)
|
|
elif opt in ('verbose', 'dry_run'): # ugh!
|
|
val = strtobool(val)
|
|
|
|
try:
|
|
setattr(self, alias or opt, val)
|
|
except ValueError as e:
|
|
raise DistutilsOptionError(e) from e
|
|
|
|
def warn_dash_deprecation(self, opt, section):
|
|
if section in (
|
|
'options.extras_require',
|
|
'options.data_files',
|
|
):
|
|
return opt
|
|
|
|
underscore_opt = opt.replace('-', '_')
|
|
commands = list(itertools.chain(
|
|
distutils.command.__all__,
|
|
self._setuptools_commands(),
|
|
))
|
|
if (
|
|
not section.startswith('options')
|
|
and section != 'metadata'
|
|
and section not in commands
|
|
):
|
|
return underscore_opt
|
|
|
|
if '-' in opt:
|
|
warnings.warn(
|
|
"Usage of dash-separated '%s' will not be supported in future "
|
|
"versions. Please use the underscore name '%s' instead"
|
|
% (opt, underscore_opt)
|
|
)
|
|
return underscore_opt
|
|
|
|
def _setuptools_commands(self):
|
|
try:
|
|
return metadata.distribution('setuptools').entry_points.names
|
|
except metadata.PackageNotFoundError:
|
|
# during bootstrapping, distribution doesn't exist
|
|
return []
|
|
|
|
def make_option_lowercase(self, opt, section):
|
|
if section != 'metadata' or opt.islower():
|
|
return opt
|
|
|
|
lowercase_opt = opt.lower()
|
|
warnings.warn(
|
|
"Usage of uppercase key '%s' in '%s' will be deprecated in future "
|
|
"versions. Please use lowercase '%s' instead"
|
|
% (opt, section, lowercase_opt)
|
|
)
|
|
return lowercase_opt
|
|
|
|
# FIXME: 'Distribution._set_command_options' is too complex (14)
|
|
def _set_command_options(self, command_obj, option_dict=None): # noqa: C901
|
|
"""
|
|
Set the options for 'command_obj' from 'option_dict'. Basically
|
|
this means copying elements of a dictionary ('option_dict') to
|
|
attributes of an instance ('command').
|
|
|
|
'command_obj' must be a Command instance. If 'option_dict' is not
|
|
supplied, uses the standard option dictionary for this command
|
|
(from 'self.command_options').
|
|
|
|
(Adopted from distutils.dist.Distribution._set_command_options)
|
|
"""
|
|
command_name = command_obj.get_command_name()
|
|
if option_dict is None:
|
|
option_dict = self.get_option_dict(command_name)
|
|
|
|
if DEBUG:
|
|
self.announce(" setting options for '%s' command:" % command_name)
|
|
for (option, (source, value)) in option_dict.items():
|
|
if DEBUG:
|
|
self.announce(" %s = %s (from %s)" % (option, value, source))
|
|
try:
|
|
bool_opts = [translate_longopt(o) for o in command_obj.boolean_options]
|
|
except AttributeError:
|
|
bool_opts = []
|
|
try:
|
|
neg_opt = command_obj.negative_opt
|
|
except AttributeError:
|
|
neg_opt = {}
|
|
|
|
try:
|
|
is_string = isinstance(value, str)
|
|
if option in neg_opt and is_string:
|
|
setattr(command_obj, neg_opt[option], not strtobool(value))
|
|
elif option in bool_opts and is_string:
|
|
setattr(command_obj, option, strtobool(value))
|
|
elif hasattr(command_obj, option):
|
|
setattr(command_obj, option, value)
|
|
else:
|
|
raise DistutilsOptionError(
|
|
"error in %s: command '%s' has no such option '%s'"
|
|
% (source, command_name, option)
|
|
)
|
|
except ValueError as e:
|
|
raise DistutilsOptionError(e) from e
|
|
|
|
def _get_project_config_files(self, filenames):
|
|
"""Add default file and split between INI and TOML"""
|
|
tomlfiles = []
|
|
standard_project_metadata = Path(self.src_root or os.curdir, "pyproject.toml")
|
|
if filenames is not None:
|
|
parts = partition(lambda f: Path(f).suffix == ".toml", filenames)
|
|
filenames = list(parts[0]) # 1st element => predicate is False
|
|
tomlfiles = list(parts[1]) # 2nd element => predicate is True
|
|
elif standard_project_metadata.exists():
|
|
tomlfiles = [standard_project_metadata]
|
|
return filenames, tomlfiles
|
|
|
|
def parse_config_files(self, filenames=None, ignore_option_errors=False):
|
|
"""Parses configuration files from various levels
|
|
and loads configuration.
|
|
"""
|
|
inifiles, tomlfiles = self._get_project_config_files(filenames)
|
|
|
|
self._parse_config_files(filenames=inifiles)
|
|
|
|
setupcfg.parse_configuration(
|
|
self, self.command_options, ignore_option_errors=ignore_option_errors
|
|
)
|
|
for filename in tomlfiles:
|
|
pyprojecttoml.apply_configuration(self, filename, ignore_option_errors)
|
|
|
|
self._finalize_requires()
|
|
self._finalize_license_files()
|
|
|
|
def fetch_build_eggs(self, requires):
|
|
"""Resolve pre-setup requirements"""
|
|
resolved_dists = pkg_resources.working_set.resolve(
|
|
_reqs.parse(requires),
|
|
installer=self.fetch_build_egg,
|
|
replace_conflicting=True,
|
|
)
|
|
for dist in resolved_dists:
|
|
pkg_resources.working_set.add(dist, replace=True)
|
|
return resolved_dists
|
|
|
|
def finalize_options(self):
|
|
"""
|
|
Allow plugins to apply arbitrary operations to the
|
|
distribution. Each hook may optionally define a 'order'
|
|
to influence the order of execution. Smaller numbers
|
|
go first and the default is 0.
|
|
"""
|
|
group = 'setuptools.finalize_distribution_options'
|
|
|
|
def by_order(hook):
|
|
return getattr(hook, 'order', 0)
|
|
|
|
defined = metadata.entry_points(group=group)
|
|
filtered = itertools.filterfalse(self._removed, defined)
|
|
loaded = map(lambda e: e.load(), filtered)
|
|
for ep in sorted(loaded, key=by_order):
|
|
ep(self)
|
|
|
|
@staticmethod
|
|
def _removed(ep):
|
|
"""
|
|
When removing an entry point, if metadata is loaded
|
|
from an older version of Setuptools, that removed
|
|
entry point will attempt to be loaded and will fail.
|
|
See #2765 for more details.
|
|
"""
|
|
removed = {
|
|
# removed 2021-09-05
|
|
'2to3_doctests',
|
|
}
|
|
return ep.name in removed
|
|
|
|
def _finalize_setup_keywords(self):
|
|
for ep in metadata.entry_points(group='distutils.setup_keywords'):
|
|
value = getattr(self, ep.name, None)
|
|
if value is not None:
|
|
self._install_dependencies(ep)
|
|
ep.load()(self, ep.name, value)
|
|
|
|
def _install_dependencies(self, ep):
|
|
"""
|
|
Given an entry point, ensure that any declared extras for
|
|
its distribution are installed.
|
|
"""
|
|
for req in nspektr.missing(ep):
|
|
# fetch_build_egg expects pkg_resources.Requirement
|
|
self.fetch_build_egg(pkg_resources.Requirement(str(req)))
|
|
|
|
def get_egg_cache_dir(self):
|
|
egg_cache_dir = os.path.join(os.curdir, '.eggs')
|
|
if not os.path.exists(egg_cache_dir):
|
|
os.mkdir(egg_cache_dir)
|
|
windows_support.hide_file(egg_cache_dir)
|
|
readme_txt_filename = os.path.join(egg_cache_dir, 'README.txt')
|
|
with open(readme_txt_filename, 'w') as f:
|
|
f.write(
|
|
'This directory contains eggs that were downloaded '
|
|
'by setuptools to build, test, and run plug-ins.\n\n'
|
|
)
|
|
f.write(
|
|
'This directory caches those eggs to prevent '
|
|
'repeated downloads.\n\n'
|
|
)
|
|
f.write('However, it is safe to delete this directory.\n\n')
|
|
|
|
return egg_cache_dir
|
|
|
|
def fetch_build_egg(self, req):
|
|
"""Fetch an egg needed for building"""
|
|
from setuptools.installer import fetch_build_egg
|
|
|
|
return fetch_build_egg(self, req)
|
|
|
|
def get_command_class(self, command):
|
|
"""Pluggable version of get_command_class()"""
|
|
if command in self.cmdclass:
|
|
return self.cmdclass[command]
|
|
|
|
eps = metadata.entry_points(group='distutils.commands', name=command)
|
|
for ep in eps:
|
|
self._install_dependencies(ep)
|
|
self.cmdclass[command] = cmdclass = ep.load()
|
|
return cmdclass
|
|
else:
|
|
return _Distribution.get_command_class(self, command)
|
|
|
|
def print_commands(self):
|
|
for ep in metadata.entry_points(group='distutils.commands'):
|
|
if ep.name not in self.cmdclass:
|
|
cmdclass = ep.load()
|
|
self.cmdclass[ep.name] = cmdclass
|
|
return _Distribution.print_commands(self)
|
|
|
|
def get_command_list(self):
|
|
for ep in metadata.entry_points(group='distutils.commands'):
|
|
if ep.name not in self.cmdclass:
|
|
cmdclass = ep.load()
|
|
self.cmdclass[ep.name] = cmdclass
|
|
return _Distribution.get_command_list(self)
|
|
|
|
def include(self, **attrs):
|
|
"""Add items to distribution that are named in keyword arguments
|
|
|
|
For example, 'dist.include(py_modules=["x"])' would add 'x' to
|
|
the distribution's 'py_modules' attribute, if it was not already
|
|
there.
|
|
|
|
Currently, this method only supports inclusion for attributes that are
|
|
lists or tuples. If you need to add support for adding to other
|
|
attributes in this or a subclass, you can add an '_include_X' method,
|
|
where 'X' is the name of the attribute. The method will be called with
|
|
the value passed to 'include()'. So, 'dist.include(foo={"bar":"baz"})'
|
|
will try to call 'dist._include_foo({"bar":"baz"})', which can then
|
|
handle whatever special inclusion logic is needed.
|
|
"""
|
|
for k, v in attrs.items():
|
|
include = getattr(self, '_include_' + k, None)
|
|
if include:
|
|
include(v)
|
|
else:
|
|
self._include_misc(k, v)
|
|
|
|
def exclude_package(self, package):
|
|
"""Remove packages, modules, and extensions in named package"""
|
|
|
|
pfx = package + '.'
|
|
if self.packages:
|
|
self.packages = [
|
|
p for p in self.packages if p != package and not p.startswith(pfx)
|
|
]
|
|
|
|
if self.py_modules:
|
|
self.py_modules = [
|
|
p for p in self.py_modules if p != package and not p.startswith(pfx)
|
|
]
|
|
|
|
if self.ext_modules:
|
|
self.ext_modules = [
|
|
p
|
|
for p in self.ext_modules
|
|
if p.name != package and not p.name.startswith(pfx)
|
|
]
|
|
|
|
def has_contents_for(self, package):
|
|
"""Return true if 'exclude_package(package)' would do something"""
|
|
|
|
pfx = package + '.'
|
|
|
|
for p in self.iter_distribution_names():
|
|
if p == package or p.startswith(pfx):
|
|
return True
|
|
|
|
def _exclude_misc(self, name, value):
|
|
"""Handle 'exclude()' for list/tuple attrs without a special handler"""
|
|
if not isinstance(value, sequence):
|
|
raise DistutilsSetupError(
|
|
"%s: setting must be a list or tuple (%r)" % (name, value)
|
|
)
|
|
try:
|
|
old = getattr(self, name)
|
|
except AttributeError as e:
|
|
raise DistutilsSetupError("%s: No such distribution setting" % name) from e
|
|
if old is not None and not isinstance(old, sequence):
|
|
raise DistutilsSetupError(
|
|
name + ": this setting cannot be changed via include/exclude"
|
|
)
|
|
elif old:
|
|
setattr(self, name, [item for item in old if item not in value])
|
|
|
|
def _include_misc(self, name, value):
|
|
"""Handle 'include()' for list/tuple attrs without a special handler"""
|
|
|
|
if not isinstance(value, sequence):
|
|
raise DistutilsSetupError("%s: setting must be a list (%r)" % (name, value))
|
|
try:
|
|
old = getattr(self, name)
|
|
except AttributeError as e:
|
|
raise DistutilsSetupError("%s: No such distribution setting" % name) from e
|
|
if old is None:
|
|
setattr(self, name, value)
|
|
elif not isinstance(old, sequence):
|
|
raise DistutilsSetupError(
|
|
name + ": this setting cannot be changed via include/exclude"
|
|
)
|
|
else:
|
|
new = [item for item in value if item not in old]
|
|
setattr(self, name, old + new)
|
|
|
|
def exclude(self, **attrs):
|
|
"""Remove items from distribution that are named in keyword arguments
|
|
|
|
For example, 'dist.exclude(py_modules=["x"])' would remove 'x' from
|
|
the distribution's 'py_modules' attribute. Excluding packages uses
|
|
the 'exclude_package()' method, so all of the package's contained
|
|
packages, modules, and extensions are also excluded.
|
|
|
|
Currently, this method only supports exclusion from attributes that are
|
|
lists or tuples. If you need to add support for excluding from other
|
|
attributes in this or a subclass, you can add an '_exclude_X' method,
|
|
where 'X' is the name of the attribute. The method will be called with
|
|
the value passed to 'exclude()'. So, 'dist.exclude(foo={"bar":"baz"})'
|
|
will try to call 'dist._exclude_foo({"bar":"baz"})', which can then
|
|
handle whatever special exclusion logic is needed.
|
|
"""
|
|
for k, v in attrs.items():
|
|
exclude = getattr(self, '_exclude_' + k, None)
|
|
if exclude:
|
|
exclude(v)
|
|
else:
|
|
self._exclude_misc(k, v)
|
|
|
|
def _exclude_packages(self, packages):
|
|
if not isinstance(packages, sequence):
|
|
raise DistutilsSetupError(
|
|
"packages: setting must be a list or tuple (%r)" % (packages,)
|
|
)
|
|
list(map(self.exclude_package, packages))
|
|
|
|
def _parse_command_opts(self, parser, args):
|
|
# Remove --with-X/--without-X options when processing command args
|
|
self.global_options = self.__class__.global_options
|
|
self.negative_opt = self.__class__.negative_opt
|
|
|
|
# First, expand any aliases
|
|
command = args[0]
|
|
aliases = self.get_option_dict('aliases')
|
|
while command in aliases:
|
|
src, alias = aliases[command]
|
|
del aliases[command] # ensure each alias can expand only once!
|
|
import shlex
|
|
|
|
args[:1] = shlex.split(alias, True)
|
|
command = args[0]
|
|
|
|
nargs = _Distribution._parse_command_opts(self, parser, args)
|
|
|
|
# Handle commands that want to consume all remaining arguments
|
|
cmd_class = self.get_command_class(command)
|
|
if getattr(cmd_class, 'command_consumes_arguments', None):
|
|
self.get_option_dict(command)['args'] = ("command line", nargs)
|
|
if nargs is not None:
|
|
return []
|
|
|
|
return nargs
|
|
|
|
def get_cmdline_options(self):
|
|
"""Return a '{cmd: {opt:val}}' map of all command-line options
|
|
|
|
Option names are all long, but do not include the leading '--', and
|
|
contain dashes rather than underscores. If the option doesn't take
|
|
an argument (e.g. '--quiet'), the 'val' is 'None'.
|
|
|
|
Note that options provided by config files are intentionally excluded.
|
|
"""
|
|
|
|
d = {}
|
|
|
|
for cmd, opts in self.command_options.items():
|
|
|
|
for opt, (src, val) in opts.items():
|
|
|
|
if src != "command line":
|
|
continue
|
|
|
|
opt = opt.replace('_', '-')
|
|
|
|
if val == 0:
|
|
cmdobj = self.get_command_obj(cmd)
|
|
neg_opt = self.negative_opt.copy()
|
|
neg_opt.update(getattr(cmdobj, 'negative_opt', {}))
|
|
for neg, pos in neg_opt.items():
|
|
if pos == opt:
|
|
opt = neg
|
|
val = None
|
|
break
|
|
else:
|
|
raise AssertionError("Shouldn't be able to get here")
|
|
|
|
elif val == 1:
|
|
val = None
|
|
|
|
d.setdefault(cmd, {})[opt] = val
|
|
|
|
return d
|
|
|
|
def iter_distribution_names(self):
|
|
"""Yield all packages, modules, and extension names in distribution"""
|
|
|
|
for pkg in self.packages or ():
|
|
yield pkg
|
|
|
|
for module in self.py_modules or ():
|
|
yield module
|
|
|
|
for ext in self.ext_modules or ():
|
|
if isinstance(ext, tuple):
|
|
name, buildinfo = ext
|
|
else:
|
|
name = ext.name
|
|
if name.endswith('module'):
|
|
name = name[:-6]
|
|
yield name
|
|
|
|
def handle_display_options(self, option_order):
|
|
"""If there were any non-global "display-only" options
|
|
(--help-commands or the metadata display options) on the command
|
|
line, display the requested info and return true; else return
|
|
false.
|
|
"""
|
|
import sys
|
|
|
|
if self.help_commands:
|
|
return _Distribution.handle_display_options(self, option_order)
|
|
|
|
# Stdout may be StringIO (e.g. in tests)
|
|
if not isinstance(sys.stdout, io.TextIOWrapper):
|
|
return _Distribution.handle_display_options(self, option_order)
|
|
|
|
# Don't wrap stdout if utf-8 is already the encoding. Provides
|
|
# workaround for #334.
|
|
if sys.stdout.encoding.lower() in ('utf-8', 'utf8'):
|
|
return _Distribution.handle_display_options(self, option_order)
|
|
|
|
# Print metadata in UTF-8 no matter the platform
|
|
encoding = sys.stdout.encoding
|
|
errors = sys.stdout.errors
|
|
newline = sys.platform != 'win32' and '\n' or None
|
|
line_buffering = sys.stdout.line_buffering
|
|
|
|
sys.stdout = io.TextIOWrapper(
|
|
sys.stdout.detach(), 'utf-8', errors, newline, line_buffering
|
|
)
|
|
try:
|
|
return _Distribution.handle_display_options(self, option_order)
|
|
finally:
|
|
sys.stdout = io.TextIOWrapper(
|
|
sys.stdout.detach(), encoding, errors, newline, line_buffering
|
|
)
|
|
|
|
def run_command(self, command):
|
|
self.set_defaults()
|
|
# Postpone defaults until all explicit configuration is considered
|
|
# (setup() args, config files, command line and plugins)
|
|
|
|
super().run_command(command)
|
|
|
|
|
|
class DistDeprecationWarning(SetuptoolsDeprecationWarning):
|
|
"""Class for warning about deprecations in dist in
|
|
setuptools. Not ignored by default, unlike DeprecationWarning."""
|