forked from brl/citadel
720 lines
30 KiB
Python
720 lines
30 KiB
Python
|
# Recipe creation tool - create build system handler for python
|
||
|
#
|
||
|
# Copyright (C) 2015 Mentor Graphics Corporation
|
||
|
#
|
||
|
# This program is free software; you can redistribute it and/or modify
|
||
|
# it under the terms of the GNU General Public License version 2 as
|
||
|
# published by the Free Software Foundation.
|
||
|
#
|
||
|
# This program is distributed in the hope that it will be useful,
|
||
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||
|
# GNU General Public License for more details.
|
||
|
#
|
||
|
# You should have received a copy of the GNU General Public License along
|
||
|
# with this program; if not, write to the Free Software Foundation, Inc.,
|
||
|
# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
|
||
|
|
||
|
import ast
|
||
|
import codecs
|
||
|
import collections
|
||
|
import distutils.command.build_py
|
||
|
import email
|
||
|
import imp
|
||
|
import glob
|
||
|
import itertools
|
||
|
import logging
|
||
|
import os
|
||
|
import re
|
||
|
import sys
|
||
|
import subprocess
|
||
|
from recipetool.create import RecipeHandler
|
||
|
|
||
|
logger = logging.getLogger('recipetool')
|
||
|
|
||
|
tinfoil = None
|
||
|
|
||
|
|
||
|
def tinfoil_init(instance):
|
||
|
global tinfoil
|
||
|
tinfoil = instance
|
||
|
|
||
|
|
||
|
class PythonRecipeHandler(RecipeHandler):
|
||
|
base_pkgdeps = ['python-core']
|
||
|
excluded_pkgdeps = ['python-dbg']
|
||
|
# os.path is provided by python-core
|
||
|
assume_provided = ['builtins', 'os.path']
|
||
|
# Assumes that the host python builtin_module_names is sane for target too
|
||
|
assume_provided = assume_provided + list(sys.builtin_module_names)
|
||
|
|
||
|
bbvar_map = {
|
||
|
'Name': 'PN',
|
||
|
'Version': 'PV',
|
||
|
'Home-page': 'HOMEPAGE',
|
||
|
'Summary': 'SUMMARY',
|
||
|
'Description': 'DESCRIPTION',
|
||
|
'License': 'LICENSE',
|
||
|
'Requires': 'RDEPENDS_${PN}',
|
||
|
'Provides': 'RPROVIDES_${PN}',
|
||
|
'Obsoletes': 'RREPLACES_${PN}',
|
||
|
}
|
||
|
# PN/PV are already set by recipetool core & desc can be extremely long
|
||
|
excluded_fields = [
|
||
|
'Description',
|
||
|
]
|
||
|
setup_parse_map = {
|
||
|
'Url': 'Home-page',
|
||
|
'Classifiers': 'Classifier',
|
||
|
'Description': 'Summary',
|
||
|
}
|
||
|
setuparg_map = {
|
||
|
'Home-page': 'url',
|
||
|
'Classifier': 'classifiers',
|
||
|
'Summary': 'description',
|
||
|
'Description': 'long-description',
|
||
|
}
|
||
|
# Values which are lists, used by the setup.py argument based metadata
|
||
|
# extraction method, to determine how to process the setup.py output.
|
||
|
setuparg_list_fields = [
|
||
|
'Classifier',
|
||
|
'Requires',
|
||
|
'Provides',
|
||
|
'Obsoletes',
|
||
|
'Platform',
|
||
|
'Supported-Platform',
|
||
|
]
|
||
|
setuparg_multi_line_values = ['Description']
|
||
|
replacements = [
|
||
|
('License', r' +$', ''),
|
||
|
('License', r'^ +', ''),
|
||
|
('License', r' ', '-'),
|
||
|
('License', r'^GNU-', ''),
|
||
|
('License', r'-[Ll]icen[cs]e(,?-[Vv]ersion)?', ''),
|
||
|
('License', r'^UNKNOWN$', ''),
|
||
|
|
||
|
# Remove currently unhandled version numbers from these variables
|
||
|
('Requires', r' *\([^)]*\)', ''),
|
||
|
('Provides', r' *\([^)]*\)', ''),
|
||
|
('Obsoletes', r' *\([^)]*\)', ''),
|
||
|
('Install-requires', r'^([^><= ]+).*', r'\1'),
|
||
|
('Extras-require', r'^([^><= ]+).*', r'\1'),
|
||
|
('Tests-require', r'^([^><= ]+).*', r'\1'),
|
||
|
|
||
|
# Remove unhandled dependency on particular features (e.g. foo[PDF])
|
||
|
('Install-requires', r'\[[^\]]+\]$', ''),
|
||
|
]
|
||
|
|
||
|
classifier_license_map = {
|
||
|
'License :: OSI Approved :: Academic Free License (AFL)': 'AFL',
|
||
|
'License :: OSI Approved :: Apache Software License': 'Apache',
|
||
|
'License :: OSI Approved :: Apple Public Source License': 'APSL',
|
||
|
'License :: OSI Approved :: Artistic License': 'Artistic',
|
||
|
'License :: OSI Approved :: Attribution Assurance License': 'AAL',
|
||
|
'License :: OSI Approved :: BSD License': 'BSD',
|
||
|
'License :: OSI Approved :: Common Public License': 'CPL',
|
||
|
'License :: OSI Approved :: Eiffel Forum License': 'EFL',
|
||
|
'License :: OSI Approved :: European Union Public Licence 1.0 (EUPL 1.0)': 'EUPL-1.0',
|
||
|
'License :: OSI Approved :: European Union Public Licence 1.1 (EUPL 1.1)': 'EUPL-1.1',
|
||
|
'License :: OSI Approved :: GNU Affero General Public License v3 or later (AGPLv3+)': 'AGPL-3.0+',
|
||
|
'License :: OSI Approved :: GNU Affero General Public License v3': 'AGPL-3.0',
|
||
|
'License :: OSI Approved :: GNU Free Documentation License (FDL)': 'GFDL',
|
||
|
'License :: OSI Approved :: GNU General Public License (GPL)': 'GPL',
|
||
|
'License :: OSI Approved :: GNU General Public License v2 (GPLv2)': 'GPL-2.0',
|
||
|
'License :: OSI Approved :: GNU General Public License v2 or later (GPLv2+)': 'GPL-2.0+',
|
||
|
'License :: OSI Approved :: GNU General Public License v3 (GPLv3)': 'GPL-3.0',
|
||
|
'License :: OSI Approved :: GNU General Public License v3 or later (GPLv3+)': 'GPL-3.0+',
|
||
|
'License :: OSI Approved :: GNU Lesser General Public License v2 (LGPLv2)': 'LGPL-2.0',
|
||
|
'License :: OSI Approved :: GNU Lesser General Public License v2 or later (LGPLv2+)': 'LGPL-2.0+',
|
||
|
'License :: OSI Approved :: GNU Lesser General Public License v3 (LGPLv3)': 'LGPL-3.0',
|
||
|
'License :: OSI Approved :: GNU Lesser General Public License v3 or later (LGPLv3+)': 'LGPL-3.0+',
|
||
|
'License :: OSI Approved :: GNU Library or Lesser General Public License (LGPL)': 'LGPL',
|
||
|
'License :: OSI Approved :: IBM Public License': 'IPL',
|
||
|
'License :: OSI Approved :: ISC License (ISCL)': 'ISC',
|
||
|
'License :: OSI Approved :: Intel Open Source License': 'Intel',
|
||
|
'License :: OSI Approved :: Jabber Open Source License': 'Jabber',
|
||
|
'License :: OSI Approved :: MIT License': 'MIT',
|
||
|
'License :: OSI Approved :: MITRE Collaborative Virtual Workspace License (CVW)': 'CVWL',
|
||
|
'License :: OSI Approved :: Motosoto License': 'Motosoto',
|
||
|
'License :: OSI Approved :: Mozilla Public License 1.0 (MPL)': 'MPL-1.0',
|
||
|
'License :: OSI Approved :: Mozilla Public License 1.1 (MPL 1.1)': 'MPL-1.1',
|
||
|
'License :: OSI Approved :: Mozilla Public License 2.0 (MPL 2.0)': 'MPL-2.0',
|
||
|
'License :: OSI Approved :: Nethack General Public License': 'NGPL',
|
||
|
'License :: OSI Approved :: Nokia Open Source License': 'Nokia',
|
||
|
'License :: OSI Approved :: Open Group Test Suite License': 'OGTSL',
|
||
|
'License :: OSI Approved :: Python License (CNRI Python License)': 'CNRI-Python',
|
||
|
'License :: OSI Approved :: Python Software Foundation License': 'PSF',
|
||
|
'License :: OSI Approved :: Qt Public License (QPL)': 'QPL',
|
||
|
'License :: OSI Approved :: Ricoh Source Code Public License': 'RSCPL',
|
||
|
'License :: OSI Approved :: Sleepycat License': 'Sleepycat',
|
||
|
'License :: OSI Approved :: Sun Industry Standards Source License (SISSL)': '-- Sun Industry Standards Source License (SISSL)',
|
||
|
'License :: OSI Approved :: Sun Public License': 'SPL',
|
||
|
'License :: OSI Approved :: University of Illinois/NCSA Open Source License': 'NCSA',
|
||
|
'License :: OSI Approved :: Vovida Software License 1.0': 'VSL-1.0',
|
||
|
'License :: OSI Approved :: W3C License': 'W3C',
|
||
|
'License :: OSI Approved :: X.Net License': 'Xnet',
|
||
|
'License :: OSI Approved :: Zope Public License': 'ZPL',
|
||
|
'License :: OSI Approved :: zlib/libpng License': 'Zlib',
|
||
|
}
|
||
|
|
||
|
def __init__(self):
|
||
|
pass
|
||
|
|
||
|
def process(self, srctree, classes, lines_before, lines_after, handled, extravalues):
|
||
|
if 'buildsystem' in handled:
|
||
|
return False
|
||
|
|
||
|
if not RecipeHandler.checkfiles(srctree, ['setup.py']):
|
||
|
return
|
||
|
|
||
|
# setup.py is always parsed to get at certain required information, such as
|
||
|
# distutils vs setuptools
|
||
|
#
|
||
|
# If egg info is available, we use it for both its PKG-INFO metadata
|
||
|
# and for its requires.txt for install_requires.
|
||
|
# If PKG-INFO is available but no egg info is, we use that for metadata in preference to
|
||
|
# the parsed setup.py, but use the install_requires info from the
|
||
|
# parsed setup.py.
|
||
|
|
||
|
setupscript = os.path.join(srctree, 'setup.py')
|
||
|
try:
|
||
|
setup_info, uses_setuptools, setup_non_literals, extensions = self.parse_setup_py(setupscript)
|
||
|
except Exception:
|
||
|
logger.exception("Failed to parse setup.py")
|
||
|
setup_info, uses_setuptools, setup_non_literals, extensions = {}, True, [], []
|
||
|
|
||
|
egginfo = glob.glob(os.path.join(srctree, '*.egg-info'))
|
||
|
if egginfo:
|
||
|
info = self.get_pkginfo(os.path.join(egginfo[0], 'PKG-INFO'))
|
||
|
requires_txt = os.path.join(egginfo[0], 'requires.txt')
|
||
|
if os.path.exists(requires_txt):
|
||
|
with codecs.open(requires_txt) as f:
|
||
|
inst_req = []
|
||
|
extras_req = collections.defaultdict(list)
|
||
|
current_feature = None
|
||
|
for line in f.readlines():
|
||
|
line = line.rstrip()
|
||
|
if not line:
|
||
|
continue
|
||
|
|
||
|
if line.startswith('['):
|
||
|
current_feature = line[1:-1]
|
||
|
elif current_feature:
|
||
|
extras_req[current_feature].append(line)
|
||
|
else:
|
||
|
inst_req.append(line)
|
||
|
info['Install-requires'] = inst_req
|
||
|
info['Extras-require'] = extras_req
|
||
|
elif RecipeHandler.checkfiles(srctree, ['PKG-INFO']):
|
||
|
info = self.get_pkginfo(os.path.join(srctree, 'PKG-INFO'))
|
||
|
|
||
|
if setup_info:
|
||
|
if 'Install-requires' in setup_info:
|
||
|
info['Install-requires'] = setup_info['Install-requires']
|
||
|
if 'Extras-require' in setup_info:
|
||
|
info['Extras-require'] = setup_info['Extras-require']
|
||
|
else:
|
||
|
if setup_info:
|
||
|
info = setup_info
|
||
|
else:
|
||
|
info = self.get_setup_args_info(setupscript)
|
||
|
|
||
|
# Grab the license value before applying replacements
|
||
|
license_str = info.get('License', '').strip()
|
||
|
|
||
|
self.apply_info_replacements(info)
|
||
|
|
||
|
if uses_setuptools:
|
||
|
classes.append('setuptools')
|
||
|
else:
|
||
|
classes.append('distutils')
|
||
|
|
||
|
if license_str:
|
||
|
for i, line in enumerate(lines_before):
|
||
|
if line.startswith('LICENSE = '):
|
||
|
lines_before.insert(i, '# NOTE: License in setup.py/PKGINFO is: %s' % license_str)
|
||
|
break
|
||
|
|
||
|
if 'Classifier' in info:
|
||
|
existing_licenses = info.get('License', '')
|
||
|
licenses = []
|
||
|
for classifier in info['Classifier']:
|
||
|
if classifier in self.classifier_license_map:
|
||
|
license = self.classifier_license_map[classifier]
|
||
|
if license == 'Apache' and 'Apache-2.0' in existing_licenses:
|
||
|
license = 'Apache-2.0'
|
||
|
elif license == 'GPL':
|
||
|
if 'GPL-2.0' in existing_licenses or 'GPLv2' in existing_licenses:
|
||
|
license = 'GPL-2.0'
|
||
|
elif 'GPL-3.0' in existing_licenses or 'GPLv3' in existing_licenses:
|
||
|
license = 'GPL-3.0'
|
||
|
elif license == 'LGPL':
|
||
|
if 'LGPL-2.1' in existing_licenses or 'LGPLv2.1' in existing_licenses:
|
||
|
license = 'LGPL-2.1'
|
||
|
elif 'LGPL-2.0' in existing_licenses or 'LGPLv2' in existing_licenses:
|
||
|
license = 'LGPL-2.0'
|
||
|
elif 'LGPL-3.0' in existing_licenses or 'LGPLv3' in existing_licenses:
|
||
|
license = 'LGPL-3.0'
|
||
|
licenses.append(license)
|
||
|
|
||
|
if licenses:
|
||
|
info['License'] = ' & '.join(licenses)
|
||
|
|
||
|
# Map PKG-INFO & setup.py fields to bitbake variables
|
||
|
for field, values in info.items():
|
||
|
if field in self.excluded_fields:
|
||
|
continue
|
||
|
|
||
|
if field not in self.bbvar_map:
|
||
|
continue
|
||
|
|
||
|
if isinstance(values, str):
|
||
|
value = values
|
||
|
else:
|
||
|
value = ' '.join(str(v) for v in values if v)
|
||
|
|
||
|
bbvar = self.bbvar_map[field]
|
||
|
if bbvar not in extravalues and value:
|
||
|
extravalues[bbvar] = value
|
||
|
|
||
|
mapped_deps, unmapped_deps = self.scan_setup_python_deps(srctree, setup_info, setup_non_literals)
|
||
|
|
||
|
extras_req = set()
|
||
|
if 'Extras-require' in info:
|
||
|
extras_req = info['Extras-require']
|
||
|
if extras_req:
|
||
|
lines_after.append('# The following configs & dependencies are from setuptools extras_require.')
|
||
|
lines_after.append('# These dependencies are optional, hence can be controlled via PACKAGECONFIG.')
|
||
|
lines_after.append('# The upstream names may not correspond exactly to bitbake package names.')
|
||
|
lines_after.append('#')
|
||
|
lines_after.append('# Uncomment this line to enable all the optional features.')
|
||
|
lines_after.append('#PACKAGECONFIG ?= "{}"'.format(' '.join(k.lower() for k in extras_req)))
|
||
|
for feature, feature_reqs in extras_req.items():
|
||
|
unmapped_deps.difference_update(feature_reqs)
|
||
|
|
||
|
feature_req_deps = ('python-' + r.replace('.', '-').lower() for r in sorted(feature_reqs))
|
||
|
lines_after.append('PACKAGECONFIG[{}] = ",,,{}"'.format(feature.lower(), ' '.join(feature_req_deps)))
|
||
|
|
||
|
inst_reqs = set()
|
||
|
if 'Install-requires' in info:
|
||
|
if extras_req:
|
||
|
lines_after.append('')
|
||
|
inst_reqs = info['Install-requires']
|
||
|
if inst_reqs:
|
||
|
unmapped_deps.difference_update(inst_reqs)
|
||
|
|
||
|
inst_req_deps = ('python-' + r.replace('.', '-').lower() for r in sorted(inst_reqs))
|
||
|
lines_after.append('# WARNING: the following rdepends are from setuptools install_requires. These')
|
||
|
lines_after.append('# upstream names may not correspond exactly to bitbake package names.')
|
||
|
lines_after.append('RDEPENDS_${{PN}} += "{}"'.format(' '.join(inst_req_deps)))
|
||
|
|
||
|
if mapped_deps:
|
||
|
name = info.get('Name')
|
||
|
if name and name[0] in mapped_deps:
|
||
|
# Attempt to avoid self-reference
|
||
|
mapped_deps.remove(name[0])
|
||
|
mapped_deps -= set(self.excluded_pkgdeps)
|
||
|
if inst_reqs or extras_req:
|
||
|
lines_after.append('')
|
||
|
lines_after.append('# WARNING: the following rdepends are determined through basic analysis of the')
|
||
|
lines_after.append('# python sources, and might not be 100% accurate.')
|
||
|
lines_after.append('RDEPENDS_${{PN}} += "{}"'.format(' '.join(sorted(mapped_deps))))
|
||
|
|
||
|
unmapped_deps -= set(extensions)
|
||
|
unmapped_deps -= set(self.assume_provided)
|
||
|
if unmapped_deps:
|
||
|
if mapped_deps:
|
||
|
lines_after.append('')
|
||
|
lines_after.append('# WARNING: We were unable to map the following python package/module')
|
||
|
lines_after.append('# dependencies to the bitbake packages which include them:')
|
||
|
lines_after.extend('# {}'.format(d) for d in sorted(unmapped_deps))
|
||
|
|
||
|
handled.append('buildsystem')
|
||
|
|
||
|
def get_pkginfo(self, pkginfo_fn):
|
||
|
msg = email.message_from_file(open(pkginfo_fn, 'r'))
|
||
|
msginfo = {}
|
||
|
for field in msg.keys():
|
||
|
values = msg.get_all(field)
|
||
|
if len(values) == 1:
|
||
|
msginfo[field] = values[0]
|
||
|
else:
|
||
|
msginfo[field] = values
|
||
|
return msginfo
|
||
|
|
||
|
def parse_setup_py(self, setupscript='./setup.py'):
|
||
|
with codecs.open(setupscript) as f:
|
||
|
info, imported_modules, non_literals, extensions = gather_setup_info(f)
|
||
|
|
||
|
def _map(key):
|
||
|
key = key.replace('_', '-')
|
||
|
key = key[0].upper() + key[1:]
|
||
|
if key in self.setup_parse_map:
|
||
|
key = self.setup_parse_map[key]
|
||
|
return key
|
||
|
|
||
|
# Naive mapping of setup() arguments to PKG-INFO field names
|
||
|
for d in [info, non_literals]:
|
||
|
for key, value in list(d.items()):
|
||
|
if key is None:
|
||
|
continue
|
||
|
new_key = _map(key)
|
||
|
if new_key != key:
|
||
|
del d[key]
|
||
|
d[new_key] = value
|
||
|
|
||
|
return info, 'setuptools' in imported_modules, non_literals, extensions
|
||
|
|
||
|
def get_setup_args_info(self, setupscript='./setup.py'):
|
||
|
cmd = ['python', setupscript]
|
||
|
info = {}
|
||
|
keys = set(self.bbvar_map.keys())
|
||
|
keys |= set(self.setuparg_list_fields)
|
||
|
keys |= set(self.setuparg_multi_line_values)
|
||
|
grouped_keys = itertools.groupby(keys, lambda k: (k in self.setuparg_list_fields, k in self.setuparg_multi_line_values))
|
||
|
for index, keys in grouped_keys:
|
||
|
if index == (True, False):
|
||
|
# Splitlines output for each arg as a list value
|
||
|
for key in keys:
|
||
|
arg = self.setuparg_map.get(key, key.lower())
|
||
|
try:
|
||
|
arg_info = self.run_command(cmd + ['--' + arg], cwd=os.path.dirname(setupscript))
|
||
|
except (OSError, subprocess.CalledProcessError):
|
||
|
pass
|
||
|
else:
|
||
|
info[key] = [l.rstrip() for l in arg_info.splitlines()]
|
||
|
elif index == (False, True):
|
||
|
# Entire output for each arg
|
||
|
for key in keys:
|
||
|
arg = self.setuparg_map.get(key, key.lower())
|
||
|
try:
|
||
|
arg_info = self.run_command(cmd + ['--' + arg], cwd=os.path.dirname(setupscript))
|
||
|
except (OSError, subprocess.CalledProcessError):
|
||
|
pass
|
||
|
else:
|
||
|
info[key] = arg_info
|
||
|
else:
|
||
|
info.update(self.get_setup_byline(list(keys), setupscript))
|
||
|
return info
|
||
|
|
||
|
def get_setup_byline(self, fields, setupscript='./setup.py'):
|
||
|
info = {}
|
||
|
|
||
|
cmd = ['python', setupscript]
|
||
|
cmd.extend('--' + self.setuparg_map.get(f, f.lower()) for f in fields)
|
||
|
try:
|
||
|
info_lines = self.run_command(cmd, cwd=os.path.dirname(setupscript)).splitlines()
|
||
|
except (OSError, subprocess.CalledProcessError):
|
||
|
pass
|
||
|
else:
|
||
|
if len(fields) != len(info_lines):
|
||
|
logger.error('Mismatch between setup.py output lines and number of fields')
|
||
|
sys.exit(1)
|
||
|
|
||
|
for lineno, line in enumerate(info_lines):
|
||
|
line = line.rstrip()
|
||
|
info[fields[lineno]] = line
|
||
|
return info
|
||
|
|
||
|
def apply_info_replacements(self, info):
|
||
|
for variable, search, replace in self.replacements:
|
||
|
if variable not in info:
|
||
|
continue
|
||
|
|
||
|
def replace_value(search, replace, value):
|
||
|
if replace is None:
|
||
|
if re.search(search, value):
|
||
|
return None
|
||
|
else:
|
||
|
new_value = re.sub(search, replace, value)
|
||
|
if value != new_value:
|
||
|
return new_value
|
||
|
return value
|
||
|
|
||
|
value = info[variable]
|
||
|
if isinstance(value, str):
|
||
|
new_value = replace_value(search, replace, value)
|
||
|
if new_value is None:
|
||
|
del info[variable]
|
||
|
elif new_value != value:
|
||
|
info[variable] = new_value
|
||
|
elif hasattr(value, 'items'):
|
||
|
for dkey, dvalue in list(value.items()):
|
||
|
new_list = []
|
||
|
for pos, a_value in enumerate(dvalue):
|
||
|
new_value = replace_value(search, replace, a_value)
|
||
|
if new_value is not None and new_value != value:
|
||
|
new_list.append(new_value)
|
||
|
|
||
|
if value != new_list:
|
||
|
value[dkey] = new_list
|
||
|
else:
|
||
|
new_list = []
|
||
|
for pos, a_value in enumerate(value):
|
||
|
new_value = replace_value(search, replace, a_value)
|
||
|
if new_value is not None and new_value != value:
|
||
|
new_list.append(new_value)
|
||
|
|
||
|
if value != new_list:
|
||
|
info[variable] = new_list
|
||
|
|
||
|
def scan_setup_python_deps(self, srctree, setup_info, setup_non_literals):
|
||
|
if 'Package-dir' in setup_info:
|
||
|
package_dir = setup_info['Package-dir']
|
||
|
else:
|
||
|
package_dir = {}
|
||
|
|
||
|
class PackageDir(distutils.command.build_py.build_py):
|
||
|
def __init__(self, package_dir):
|
||
|
self.package_dir = package_dir
|
||
|
|
||
|
pd = PackageDir(package_dir)
|
||
|
to_scan = []
|
||
|
if not any(v in setup_non_literals for v in ['Py-modules', 'Scripts', 'Packages']):
|
||
|
if 'Py-modules' in setup_info:
|
||
|
for module in setup_info['Py-modules']:
|
||
|
try:
|
||
|
package, module = module.rsplit('.', 1)
|
||
|
except ValueError:
|
||
|
package, module = '.', module
|
||
|
module_path = os.path.join(pd.get_package_dir(package), module + '.py')
|
||
|
to_scan.append(module_path)
|
||
|
|
||
|
if 'Packages' in setup_info:
|
||
|
for package in setup_info['Packages']:
|
||
|
to_scan.append(pd.get_package_dir(package))
|
||
|
|
||
|
if 'Scripts' in setup_info:
|
||
|
to_scan.extend(setup_info['Scripts'])
|
||
|
else:
|
||
|
logger.info("Scanning the entire source tree, as one or more of the following setup keywords are non-literal: py_modules, scripts, packages.")
|
||
|
|
||
|
if not to_scan:
|
||
|
to_scan = ['.']
|
||
|
|
||
|
logger.info("Scanning paths for packages & dependencies: %s", ', '.join(to_scan))
|
||
|
|
||
|
provided_packages = self.parse_pkgdata_for_python_packages()
|
||
|
scanned_deps = self.scan_python_dependencies([os.path.join(srctree, p) for p in to_scan])
|
||
|
mapped_deps, unmapped_deps = set(self.base_pkgdeps), set()
|
||
|
for dep in scanned_deps:
|
||
|
mapped = provided_packages.get(dep)
|
||
|
if mapped:
|
||
|
logger.debug('Mapped %s to %s' % (dep, mapped))
|
||
|
mapped_deps.add(mapped)
|
||
|
else:
|
||
|
logger.debug('Could not map %s' % dep)
|
||
|
unmapped_deps.add(dep)
|
||
|
return mapped_deps, unmapped_deps
|
||
|
|
||
|
def scan_python_dependencies(self, paths):
|
||
|
deps = set()
|
||
|
try:
|
||
|
dep_output = self.run_command(['pythondeps', '-d'] + paths)
|
||
|
except (OSError, subprocess.CalledProcessError):
|
||
|
pass
|
||
|
else:
|
||
|
for line in dep_output.splitlines():
|
||
|
line = line.rstrip()
|
||
|
dep, filename = line.split('\t', 1)
|
||
|
if filename.endswith('/setup.py'):
|
||
|
continue
|
||
|
deps.add(dep)
|
||
|
|
||
|
try:
|
||
|
provides_output = self.run_command(['pythondeps', '-p'] + paths)
|
||
|
except (OSError, subprocess.CalledProcessError):
|
||
|
pass
|
||
|
else:
|
||
|
provides_lines = (l.rstrip() for l in provides_output.splitlines())
|
||
|
provides = set(l for l in provides_lines if l and l != 'setup')
|
||
|
deps -= provides
|
||
|
|
||
|
return deps
|
||
|
|
||
|
def parse_pkgdata_for_python_packages(self):
|
||
|
suffixes = [t[0] for t in imp.get_suffixes()]
|
||
|
pkgdata_dir = tinfoil.config_data.getVar('PKGDATA_DIR')
|
||
|
|
||
|
ldata = tinfoil.config_data.createCopy()
|
||
|
bb.parse.handle('classes/python-dir.bbclass', ldata, True)
|
||
|
python_sitedir = ldata.getVar('PYTHON_SITEPACKAGES_DIR')
|
||
|
|
||
|
dynload_dir = os.path.join(os.path.dirname(python_sitedir), 'lib-dynload')
|
||
|
python_dirs = [python_sitedir + os.sep,
|
||
|
os.path.join(os.path.dirname(python_sitedir), 'dist-packages') + os.sep,
|
||
|
os.path.dirname(python_sitedir) + os.sep]
|
||
|
packages = {}
|
||
|
for pkgdatafile in glob.glob('{}/runtime/*'.format(pkgdata_dir)):
|
||
|
files_info = None
|
||
|
with open(pkgdatafile, 'r') as f:
|
||
|
for line in f.readlines():
|
||
|
field, value = line.split(': ', 1)
|
||
|
if field == 'FILES_INFO':
|
||
|
files_info = ast.literal_eval(value)
|
||
|
break
|
||
|
else:
|
||
|
continue
|
||
|
|
||
|
for fn in files_info:
|
||
|
for suffix in suffixes:
|
||
|
if fn.endswith(suffix):
|
||
|
break
|
||
|
else:
|
||
|
continue
|
||
|
|
||
|
if fn.startswith(dynload_dir + os.sep):
|
||
|
if '/.debug/' in fn:
|
||
|
continue
|
||
|
base = os.path.basename(fn)
|
||
|
provided = base.split('.', 1)[0]
|
||
|
packages[provided] = os.path.basename(pkgdatafile)
|
||
|
continue
|
||
|
|
||
|
for python_dir in python_dirs:
|
||
|
if fn.startswith(python_dir):
|
||
|
relpath = fn[len(python_dir):]
|
||
|
relstart, _, relremaining = relpath.partition(os.sep)
|
||
|
if relstart.endswith('.egg'):
|
||
|
relpath = relremaining
|
||
|
base, _ = os.path.splitext(relpath)
|
||
|
|
||
|
if '/.debug/' in base:
|
||
|
continue
|
||
|
if os.path.basename(base) == '__init__':
|
||
|
base = os.path.dirname(base)
|
||
|
base = base.replace(os.sep + os.sep, os.sep)
|
||
|
provided = base.replace(os.sep, '.')
|
||
|
packages[provided] = os.path.basename(pkgdatafile)
|
||
|
return packages
|
||
|
|
||
|
@classmethod
|
||
|
def run_command(cls, cmd, **popenargs):
|
||
|
if 'stderr' not in popenargs:
|
||
|
popenargs['stderr'] = subprocess.STDOUT
|
||
|
try:
|
||
|
return subprocess.check_output(cmd, **popenargs).decode('utf-8')
|
||
|
except OSError as exc:
|
||
|
logger.error('Unable to run `{}`: {}', ' '.join(cmd), exc)
|
||
|
raise
|
||
|
except subprocess.CalledProcessError as exc:
|
||
|
logger.error('Unable to run `{}`: {}', ' '.join(cmd), exc.output)
|
||
|
raise
|
||
|
|
||
|
|
||
|
def gather_setup_info(fileobj):
|
||
|
parsed = ast.parse(fileobj.read(), fileobj.name)
|
||
|
visitor = SetupScriptVisitor()
|
||
|
visitor.visit(parsed)
|
||
|
|
||
|
non_literals, extensions = {}, []
|
||
|
for key, value in list(visitor.keywords.items()):
|
||
|
if key == 'ext_modules':
|
||
|
if isinstance(value, list):
|
||
|
for ext in value:
|
||
|
if (isinstance(ext, ast.Call) and
|
||
|
isinstance(ext.func, ast.Name) and
|
||
|
ext.func.id == 'Extension' and
|
||
|
not has_non_literals(ext.args)):
|
||
|
extensions.append(ext.args[0])
|
||
|
elif has_non_literals(value):
|
||
|
non_literals[key] = value
|
||
|
del visitor.keywords[key]
|
||
|
|
||
|
return visitor.keywords, visitor.imported_modules, non_literals, extensions
|
||
|
|
||
|
|
||
|
class SetupScriptVisitor(ast.NodeVisitor):
|
||
|
def __init__(self):
|
||
|
ast.NodeVisitor.__init__(self)
|
||
|
self.keywords = {}
|
||
|
self.non_literals = []
|
||
|
self.imported_modules = set()
|
||
|
|
||
|
def visit_Expr(self, node):
|
||
|
if isinstance(node.value, ast.Call) and \
|
||
|
isinstance(node.value.func, ast.Name) and \
|
||
|
node.value.func.id == 'setup':
|
||
|
self.visit_setup(node.value)
|
||
|
|
||
|
def visit_setup(self, node):
|
||
|
call = LiteralAstTransform().visit(node)
|
||
|
self.keywords = call.keywords
|
||
|
for k, v in self.keywords.items():
|
||
|
if has_non_literals(v):
|
||
|
self.non_literals.append(k)
|
||
|
|
||
|
def visit_Import(self, node):
|
||
|
for alias in node.names:
|
||
|
self.imported_modules.add(alias.name)
|
||
|
|
||
|
def visit_ImportFrom(self, node):
|
||
|
self.imported_modules.add(node.module)
|
||
|
|
||
|
|
||
|
class LiteralAstTransform(ast.NodeTransformer):
|
||
|
"""Simplify the ast through evaluation of literals."""
|
||
|
excluded_fields = ['ctx']
|
||
|
|
||
|
def visit(self, node):
|
||
|
if not isinstance(node, ast.AST):
|
||
|
return node
|
||
|
else:
|
||
|
return ast.NodeTransformer.visit(self, node)
|
||
|
|
||
|
def generic_visit(self, node):
|
||
|
try:
|
||
|
return ast.literal_eval(node)
|
||
|
except ValueError:
|
||
|
for field, value in ast.iter_fields(node):
|
||
|
if field in self.excluded_fields:
|
||
|
delattr(node, field)
|
||
|
if value is None:
|
||
|
continue
|
||
|
|
||
|
if isinstance(value, list):
|
||
|
if field in ('keywords', 'kwargs'):
|
||
|
new_value = dict((kw.arg, self.visit(kw.value)) for kw in value)
|
||
|
else:
|
||
|
new_value = [self.visit(i) for i in value]
|
||
|
else:
|
||
|
new_value = self.visit(value)
|
||
|
setattr(node, field, new_value)
|
||
|
return node
|
||
|
|
||
|
def visit_Name(self, node):
|
||
|
if hasattr('__builtins__', node.id):
|
||
|
return getattr(__builtins__, node.id)
|
||
|
else:
|
||
|
return self.generic_visit(node)
|
||
|
|
||
|
def visit_Tuple(self, node):
|
||
|
return tuple(self.visit(v) for v in node.elts)
|
||
|
|
||
|
def visit_List(self, node):
|
||
|
return [self.visit(v) for v in node.elts]
|
||
|
|
||
|
def visit_Set(self, node):
|
||
|
return set(self.visit(v) for v in node.elts)
|
||
|
|
||
|
def visit_Dict(self, node):
|
||
|
keys = (self.visit(k) for k in node.keys)
|
||
|
values = (self.visit(v) for v in node.values)
|
||
|
return dict(zip(keys, values))
|
||
|
|
||
|
|
||
|
def has_non_literals(value):
|
||
|
if isinstance(value, ast.AST):
|
||
|
return True
|
||
|
elif isinstance(value, str):
|
||
|
return False
|
||
|
elif hasattr(value, 'values'):
|
||
|
return any(has_non_literals(v) for v in value.values())
|
||
|
elif hasattr(value, '__iter__'):
|
||
|
return any(has_non_literals(v) for v in value)
|
||
|
|
||
|
|
||
|
def register_recipe_handlers(handlers):
|
||
|
# We need to make sure this is ahead of the makefile fallback handler
|
||
|
handlers.append((PythonRecipeHandler(), 70))
|