aboutsummaryrefslogtreecommitdiff
path: root/pugl/waflib/extras/autowaf.py
diff options
context:
space:
mode:
Diffstat (limited to 'pugl/waflib/extras/autowaf.py')
-rw-r--r--pugl/waflib/extras/autowaf.py1452
1 files changed, 1452 insertions, 0 deletions
diff --git a/pugl/waflib/extras/autowaf.py b/pugl/waflib/extras/autowaf.py
new file mode 100644
index 0000000..5feef03
--- /dev/null
+++ b/pugl/waflib/extras/autowaf.py
@@ -0,0 +1,1452 @@
+import glob
+import os
+import subprocess
+import sys
+import time
+
+from waflib import Configure, ConfigSet, Build, Context, Logs, Options, Utils
+from waflib.TaskGen import feature, before, after
+
+global g_is_child
+g_is_child = False
+
+NONEMPTY = -10
+
+if sys.platform == 'win32':
+ lib_path_name = 'PATH'
+elif sys.platform == 'darwin':
+ lib_path_name = 'DYLD_LIBRARY_PATH'
+else:
+ lib_path_name = 'LD_LIBRARY_PATH'
+
+# Compute dependencies globally
+# import preproc
+# preproc.go_absolute = True
+
+@feature('c', 'cxx')
+@after('apply_incpaths')
+def include_config_h(self):
+ self.env.append_value('INCPATHS', self.bld.bldnode.abspath())
+
+class OptionsContext(Options.OptionsContext):
+ def __init__(self, **kwargs):
+ super(OptionsContext, self).__init__(**kwargs)
+ set_options(self)
+
+ def configuration_options(self):
+ return self.get_option_group('Configuration options')
+
+ def add_flags(self, group, flags):
+ """Tersely add flags (a dictionary of longname:desc) to a group"""
+ for name, desc in flags.items():
+ group.add_option('--' + name, action='store_true',
+ dest=name.replace('-', '_'), help=desc)
+
+def set_options(opt, debug_by_default=False):
+ "Add standard autowaf options"
+ opts = opt.get_option_group('Configuration options')
+
+ # Standard directory options
+ opts.add_option('--bindir', type='string',
+ help="executable programs [default: PREFIX/bin]")
+ opts.add_option('--configdir', type='string',
+ help="configuration data [default: PREFIX/etc]")
+ opts.add_option('--datadir', type='string',
+ help="shared data [default: PREFIX/share]")
+ opts.add_option('--includedir', type='string',
+ help="header files [default: PREFIX/include]")
+ opts.add_option('--libdir', type='string',
+ help="libraries [default: PREFIX/lib]")
+ opts.add_option('--mandir', type='string',
+ help="manual pages [default: DATADIR/man]")
+ opts.add_option('--docdir', type='string',
+ help="HTML documentation [default: DATADIR/doc]")
+
+ # Build options
+ if debug_by_default:
+ opts.add_option('--optimize', action='store_false', default=True,
+ dest='debug', help="build optimized binaries")
+ else:
+ opts.add_option('-d', '--debug', action='store_true', default=False,
+ dest='debug', help="build debuggable binaries")
+ opts.add_option('--pardebug', action='store_true', default=False,
+ dest='pardebug',
+ help="build debug libraries with D suffix")
+
+ opts.add_option('-s', '--strict', action='store_true', default=False,
+ dest='strict',
+ help="use strict compiler flags and show all warnings")
+ opts.add_option('-S', '--ultra-strict', action='store_true', default=False,
+ dest='ultra_strict',
+ help="use extremely strict compiler flags (likely noisy)")
+ opts.add_option('--docs', action='store_true', default=False, dest='docs',
+ help="build documentation (requires doxygen)")
+
+ # Test options
+ if hasattr(Context.g_module, 'test'):
+ test_opts = opt.add_option_group('Test options', '')
+ opts.add_option('-T', '--test', action='store_true', dest='build_tests',
+ help='build unit tests')
+ opts.add_option('--no-coverage', action='store_true',
+ dest='no_coverage',
+ help='do not instrument code for test coverage')
+ test_opts.add_option('--wrapper', type='string',
+ dest='test_wrapper',
+ help='command prefix for tests (e.g. valgrind)')
+ test_opts.add_option('--test-filter', type='string',
+ dest='test_filter',
+ help='regular expression for tests to run')
+
+ # Run options
+ run_opts = opt.add_option_group('Run options')
+ run_opts.add_option('--cmd', type='string', dest='cmd',
+ help='command to run from build directory')
+
+class ConfigureContext(Configure.ConfigurationContext):
+ """configures the project"""
+
+ def __init__(self, **kwargs):
+ self.line_just = 45
+ if hasattr(Context.g_module, 'line_just'):
+ self.line_just = Context.g_module.line_just
+
+ super(ConfigureContext, self).__init__(**kwargs)
+ self.run_env = ConfigSet.ConfigSet()
+ self.system_include_paths = set()
+
+ def pre_recurse(self, node):
+ if len(self.stack_path) == 1:
+ Logs.pprint('BOLD', 'Configuring %s' % node.parent.srcpath())
+ super(ConfigureContext, self).pre_recurse(node)
+
+ def store(self):
+ self.env.AUTOWAF_RUN_ENV = self.run_env.get_merged_dict()
+ for path in sorted(self.system_include_paths):
+ if 'COMPILER_CC' in self.env:
+ self.env.append_value('CFLAGS', ['-isystem', path])
+ if 'COMPILER_CXX' in self.env:
+ self.env.append_value('CXXFLAGS', ['-isystem', path])
+
+ super(ConfigureContext, self).store()
+
+ def build_path(self, path='.'):
+ """Return `path` within the build directory"""
+ return str(self.path.get_bld().find_node(path))
+
+def get_check_func(conf, lang):
+ if lang == 'c':
+ return conf.check_cc
+ elif lang == 'cxx':
+ return conf.check_cxx
+ else:
+ Logs.error("Unknown header language `%s'" % lang)
+
+def check_header(conf, lang, name, define='', mandatory=True):
+ "Check for a header"
+ check_func = get_check_func(conf, lang)
+ if define != '':
+ check_func(header_name=name,
+ define_name=define,
+ mandatory=mandatory)
+ else:
+ check_func(header_name=name, mandatory=mandatory)
+
+def check_function(conf, lang, name, **args):
+ "Check for a function"
+ header_names = Utils.to_list(args['header_name'])
+ includes = ''.join(['#include <%s>\n' % x for x in header_names])
+ fragment = '''
+%s
+int main() { return !(void(*)())(%s); }
+''' % (includes, name)
+
+ check_func = get_check_func(conf, lang)
+ args['msg'] = 'Checking for %s' % name
+ check_func(fragment=fragment, **args)
+
+def nameify(name):
+ return (name.replace('/', '_').replace('++', 'PP')
+ .replace('-', '_').replace('.', '_'))
+
+def define(conf, var_name, value):
+ conf.define(var_name, value)
+ conf.env[var_name] = value
+
+def check_pkg(conf, name, **args):
+ "Check for a package iff it hasn't been checked for yet"
+ if args['uselib_store'].lower() in conf.env['AUTOWAF_LOCAL_LIBS']:
+ return
+
+ class CheckType:
+ OPTIONAL = 1
+ MANDATORY = 2
+
+ var_name = 'CHECKED_' + nameify(args['uselib_store'])
+ check = var_name not in conf.env
+ mandatory = 'mandatory' not in args or args['mandatory']
+ if not check and 'atleast_version' in args:
+ # Re-check if version is newer than previous check
+ checked_version = conf.env['VERSION_' + name]
+ if checked_version and checked_version < args['atleast_version']:
+ check = True
+ if not check and mandatory and conf.env[var_name] == CheckType.OPTIONAL:
+ # Re-check if previous check was optional but this one is mandatory
+ check = True
+ if check:
+ found = None
+ pkg_var_name = 'PKG_' + name.replace('-', '_')
+ pkg_name = name
+ if conf.env.PARDEBUG:
+ args['mandatory'] = False # Smash mandatory arg
+ found = conf.check_cfg(package=pkg_name + 'D',
+ args="--cflags --libs", **args)
+ if found:
+ pkg_name += 'D'
+ if mandatory:
+ args['mandatory'] = True # Unsmash mandatory arg
+ if not found:
+ found = conf.check_cfg(package=pkg_name, args="--cflags --libs",
+ **args)
+ if found:
+ conf.env[pkg_var_name] = pkg_name
+ if 'atleast_version' in args:
+ conf.env['VERSION_' + name] = args['atleast_version']
+ if mandatory:
+ conf.env[var_name] = CheckType.MANDATORY
+ else:
+ conf.env[var_name] = CheckType.OPTIONAL
+
+ if not conf.env.MSVC_COMPILER and 'system' in args and args['system']:
+ conf.system_include_paths.update(
+ conf.env['INCLUDES_' + nameify(args['uselib_store'])])
+
+def normpath(path):
+ if sys.platform == 'win32':
+ return os.path.normpath(path).replace('\\', '/')
+ else:
+ return os.path.normpath(path)
+
+def configure(conf):
+ def append_cxx_flags(flags):
+ conf.env.append_value('CFLAGS', flags)
+ conf.env.append_value('CXXFLAGS', flags)
+
+ if Options.options.docs:
+ conf.load('doxygen')
+
+ try:
+ conf.load('clang_compilation_database')
+ except Exception:
+ pass
+
+ prefix = normpath(os.path.abspath(os.path.expanduser(conf.env['PREFIX'])))
+
+ conf.env['DOCS'] = Options.options.docs and conf.env.DOXYGEN
+ conf.env['DEBUG'] = Options.options.debug or Options.options.pardebug
+ conf.env['PARDEBUG'] = Options.options.pardebug
+ conf.env['PREFIX'] = prefix
+
+ def config_dir(var, opt, default):
+ if opt:
+ conf.env[var] = normpath(opt)
+ else:
+ conf.env[var] = normpath(default)
+
+ opts = Options.options
+
+ config_dir('BINDIR', opts.bindir, os.path.join(prefix, 'bin'))
+ config_dir('SYSCONFDIR', opts.configdir, os.path.join(prefix, 'etc'))
+ config_dir('DATADIR', opts.datadir, os.path.join(prefix, 'share'))
+ config_dir('INCLUDEDIR', opts.includedir, os.path.join(prefix, 'include'))
+ config_dir('LIBDIR', opts.libdir, os.path.join(prefix, 'lib'))
+
+ datadir = conf.env['DATADIR']
+ config_dir('MANDIR', opts.mandir, os.path.join(datadir, 'man'))
+ config_dir('DOCDIR', opts.docdir, os.path.join(datadir, 'doc'))
+
+ if Options.options.debug:
+ if conf.env['MSVC_COMPILER']:
+ conf.env['CFLAGS'] = ['/Od', '/Z7', '/MTd', '/FS']
+ conf.env['CXXFLAGS'] = ['/Od', '/Z7', '/MTd', '/FS']
+ conf.env['LINKFLAGS'] = ['/DEBUG', '/MANIFEST']
+ else:
+ conf.env['CFLAGS'] = ['-O0', '-g']
+ conf.env['CXXFLAGS'] = ['-O0', '-g']
+ else:
+ if conf.env['MSVC_COMPILER']:
+ append_cxx_flags(['/MD', '/FS', '/DNDEBUG'])
+ else:
+ append_cxx_flags(['-DNDEBUG'])
+
+ if conf.env.MSVC_COMPILER:
+ Options.options.no_coverage = True
+ append_cxx_flags(['/nologo',
+ '/FS',
+ '/DNDEBUG',
+ '/D_CRT_SECURE_NO_WARNINGS',
+ '/experimental:external',
+ '/external:W0',
+ '/external:anglebrackets'])
+ conf.env.append_value('LINKFLAGS', '/nologo')
+ if Options.options.strict or Options.options.ultra_strict:
+ ms_strict_flags = ['/Wall',
+ '/wd4061',
+ '/wd4200',
+ '/wd4514',
+ '/wd4571',
+ '/wd4625',
+ '/wd4626',
+ '/wd4706',
+ '/wd4710',
+ '/wd4820',
+ '/wd5026',
+ '/wd5027',
+ '/wd5045']
+ conf.env.append_value('CFLAGS', ms_strict_flags)
+ conf.env.append_value('CXXFLAGS', ms_strict_flags)
+ conf.env.append_value('CXXFLAGS', ['/EHsc'])
+ else:
+ if Options.options.ultra_strict:
+ Options.options.strict = True
+ conf.env.append_value('CFLAGS', ['-Wredundant-decls',
+ '-Wstrict-prototypes',
+ '-Wmissing-prototypes',
+ '-Wcast-qual'])
+ conf.env.append_value('CXXFLAGS', ['-Wcast-qual'])
+
+ if Options.options.strict:
+ conf.env.append_value('CFLAGS', ['-pedantic', '-Wshadow'])
+ if conf.env.DEST_OS != "darwin":
+ conf.env.append_value('LINKFLAGS', ['-Wl,--no-undefined'])
+ conf.env.append_value('CXXFLAGS', ['-Wnon-virtual-dtor',
+ '-Woverloaded-virtual'])
+ append_cxx_flags(['-Wall',
+ '-Wcast-align',
+ '-Wextra',
+ '-Wmissing-declarations',
+ '-Wno-unused-parameter',
+ '-Wstrict-overflow',
+ '-Wundef',
+ '-Wwrite-strings',
+ '-fstrict-overflow'])
+
+ # Add less universal flags after checking they work
+ extra_flags = ['-Wlogical-op',
+ '-Wsuggest-attribute=noreturn',
+ '-Wunsafe-loop-optimizations']
+ if conf.check_cc(cflags=['-Werror'] + extra_flags, mandatory=False,
+ msg="Checking for extra C warning flags"):
+ conf.env.append_value('CFLAGS', extra_flags)
+ if 'COMPILER_CXX' in conf.env:
+ if conf.check_cxx(cxxflags=['-Werror'] + extra_flags,
+ mandatory=False,
+ msg="Checking for extra C++ warning flags"):
+ conf.env.append_value('CXXFLAGS', extra_flags)
+
+ if not conf.env['MSVC_COMPILER']:
+ append_cxx_flags(['-fshow-column'])
+
+ conf.env.NO_COVERAGE = True
+ conf.env.BUILD_TESTS = False
+ try:
+ conf.env.BUILD_TESTS = Options.options.build_tests
+ conf.env.NO_COVERAGE = Options.options.no_coverage
+ if not Options.options.no_coverage:
+ # Set up unit test code coverage
+ if conf.is_defined('CLANG'):
+ for cov in [conf.env.CC[0].replace('clang', 'llvm-cov'),
+ 'llvm-cov']:
+ if conf.find_program(cov, var='LLVM_COV', mandatory=False):
+ break
+ else:
+ conf.check_cc(lib='gcov', define_name='HAVE_GCOV',
+ mandatory=False)
+ except Exception:
+ pass # Test options do not exist
+
+ # Define version in configuration
+ appname = getattr(Context.g_module, Context.APPNAME, 'noname')
+ version = getattr(Context.g_module, Context.VERSION, '0.0.0')
+ defname = appname.upper().replace('-', '_').replace('.', '_')
+ define(conf, defname + '_VERSION', version)
+
+ conf.env.prepend_value('CFLAGS', '-I' + os.path.abspath('.'))
+ conf.env.prepend_value('CXXFLAGS', '-I' + os.path.abspath('.'))
+
+def display_summary(conf, msgs=None):
+ if len(conf.stack_path) == 1:
+ display_msg(conf, "Install prefix", conf.env['PREFIX'])
+ if 'COMPILER_CC' in conf.env:
+ display_msg(conf, "C Flags", ' '.join(conf.env['CFLAGS']))
+ if 'COMPILER_CXX' in conf.env:
+ display_msg(conf, "C++ Flags", ' '.join(conf.env['CXXFLAGS']))
+ display_msg(conf, "Debuggable", bool(conf.env['DEBUG']))
+ display_msg(conf, "Build documentation", bool(conf.env['DOCS']))
+
+ if msgs is not None:
+ display_msgs(conf, msgs)
+
+def set_c_lang(conf, lang):
+ "Set a specific C language standard, like 'c99' or 'c11'"
+ if conf.env.MSVC_COMPILER:
+ # MSVC has no hope or desire to compile C99, just compile as C++
+ conf.env.append_unique('CFLAGS', ['/TP'])
+ else:
+ flag = '-std=%s' % lang
+ conf.check(cflags=['-Werror', flag],
+ msg="Checking for flag '%s'" % flag)
+ conf.env.append_unique('CFLAGS', [flag])
+
+def set_cxx_lang(conf, lang):
+ "Set a specific C++ language standard, like 'c++11', 'c++14', or 'c++17'"
+ if conf.env.MSVC_COMPILER:
+ if lang != 'c++14':
+ lang = 'c++latest'
+ conf.env.append_unique('CXXFLAGS', ['/std:%s' % lang])
+ else:
+ flag = '-std=%s' % lang
+ conf.check(cxxflags=['-Werror', flag],
+ msg="Checking for flag '%s'" % flag)
+ conf.env.append_unique('CXXFLAGS', [flag])
+
+def set_modern_c_flags(conf):
+ "Use the most modern C language available"
+ if 'COMPILER_CC' in conf.env:
+ if conf.env.MSVC_COMPILER:
+ # MSVC has no hope or desire to compile C99, just compile as C++
+ conf.env.append_unique('CFLAGS', ['/TP'])
+ else:
+ for flag in ['-std=c11', '-std=c99']:
+ if conf.check(cflags=['-Werror', flag], mandatory=False,
+ msg="Checking for flag '%s'" % flag):
+ conf.env.append_unique('CFLAGS', [flag])
+ break
+
+def set_modern_cxx_flags(conf, mandatory=False):
+ "Use the most modern C++ language available"
+ if 'COMPILER_CXX' in conf.env:
+ if conf.env.MSVC_COMPILER:
+ conf.env.append_unique('CXXFLAGS', ['/std:c++latest'])
+ else:
+ for lang in ['c++14', 'c++1y', 'c++11', 'c++0x']:
+ flag = '-std=%s' % lang
+ if conf.check(cxxflags=['-Werror', flag], mandatory=False,
+ msg="Checking for flag '%s'" % flag):
+ conf.env.append_unique('CXXFLAGS', [flag])
+ break
+
+def set_local_lib(conf, name, has_objects):
+ var_name = 'HAVE_' + nameify(name.upper())
+ define(conf, var_name, 1)
+ if has_objects:
+ if type(conf.env['AUTOWAF_LOCAL_LIBS']) != dict:
+ conf.env['AUTOWAF_LOCAL_LIBS'] = {}
+ conf.env['AUTOWAF_LOCAL_LIBS'][name.lower()] = True
+ else:
+ if type(conf.env['AUTOWAF_LOCAL_HEADERS']) != dict:
+ conf.env['AUTOWAF_LOCAL_HEADERS'] = {}
+ conf.env['AUTOWAF_LOCAL_HEADERS'][name.lower()] = True
+
+def append_property(obj, key, val):
+ if hasattr(obj, key):
+ setattr(obj, key, getattr(obj, key) + val)
+ else:
+ setattr(obj, key, val)
+
+def use_lib(bld, obj, libs):
+ abssrcdir = os.path.abspath('.')
+ libs_list = libs.split()
+ for l in libs_list:
+ in_headers = l.lower() in bld.env['AUTOWAF_LOCAL_HEADERS']
+ in_libs = l.lower() in bld.env['AUTOWAF_LOCAL_LIBS']
+ if in_libs:
+ append_property(obj, 'use', ' lib%s ' % l.lower())
+ append_property(obj, 'framework', bld.env['FRAMEWORK_' + l])
+ if in_headers or in_libs:
+ if bld.env.MSVC_COMPILER:
+ inc_flag = '/I' + os.path.join(abssrcdir, l.lower())
+ else:
+ inc_flag = '-iquote ' + os.path.join(abssrcdir, l.lower())
+ for f in ['CFLAGS', 'CXXFLAGS']:
+ if inc_flag not in bld.env[f]:
+ bld.env.prepend_value(f, inc_flag)
+ else:
+ append_property(obj, 'uselib', ' ' + l)
+
+@feature('c', 'cxx')
+@before('apply_link')
+def version_lib(self):
+ if self.env.DEST_OS == 'win32':
+ self.vnum = None # Prevent waf from automatically appending -0
+ if self.env['PARDEBUG']:
+ applicable = ['cshlib', 'cxxshlib', 'cstlib', 'cxxstlib']
+ if [x for x in applicable if x in self.features]:
+ self.target = self.target + 'D'
+
+def set_lib_env(conf, name, version):
+ "Set up environment for local library as if found via pkg-config."
+ NAME = name.upper()
+ major_ver = version.split('.')[0]
+ pkg_var_name = 'PKG_' + name.replace('-', '_') + '_' + major_ver
+ lib_name = '%s-%s' % (name, major_ver)
+ lib_path = [str(conf.path.get_bld())]
+ if conf.env.PARDEBUG:
+ lib_name += 'D'
+ conf.env[pkg_var_name] = lib_name
+ conf.env['INCLUDES_' + NAME] = ['${INCLUDEDIR}/%s-%s' % (name, major_ver)]
+ conf.env['LIBPATH_' + NAME] = lib_path
+ conf.env['LIB_' + NAME] = [lib_name]
+
+ conf.run_env.append_unique(lib_path_name, lib_path)
+ conf.define(NAME + '_VERSION', version)
+
+def display_msg(conf, msg, status=None, color=None):
+ color = 'CYAN'
+ if type(status) == bool and status:
+ color = 'GREEN'
+ status = 'yes'
+ elif type(status) == bool and not status or status == "False":
+ color = 'YELLOW'
+ status = 'no'
+ Logs.pprint('BOLD', '%s' % msg.ljust(conf.line_just), sep='')
+ Logs.pprint('BOLD', ":", sep='')
+ Logs.pprint(color, status)
+
+def display_msgs(conf, msgs):
+ for k, v in msgs.items():
+ display_msg(conf, k, v)
+
+def link_flags(env, lib):
+ return ' '.join(map(lambda x: env['LIB_ST'] % x,
+ env['LIB_' + lib]))
+
+def compile_flags(env, lib):
+ return ' '.join(map(lambda x: env['CPPPATH_ST'] % x,
+ env['INCLUDES_' + lib]))
+
+def set_recursive():
+ global g_is_child
+ g_is_child = True
+
+def is_child():
+ global g_is_child
+ return g_is_child
+
+def build_pc(bld, name, version, version_suffix, libs, subst_dict={}):
+ """Build a pkg-config file for a library.
+
+ name -- uppercase variable name (e.g. 'SOMENAME')
+ version -- version string (e.g. '1.2.3')
+ version_suffix -- name version suffix (e.g. '2')
+ libs -- string/list of dependencies (e.g. 'LIBFOO GLIB')
+ """
+
+ pkg_prefix = bld.env['PREFIX']
+ if len(pkg_prefix) > 1 and pkg_prefix[-1] == '/':
+ pkg_prefix = pkg_prefix[:-1]
+
+ target = name.lower()
+ if version_suffix != '':
+ target += '-' + version_suffix
+
+ if bld.env['PARDEBUG']:
+ target += 'D'
+
+ target += '.pc'
+
+ libdir = bld.env['LIBDIR']
+ if libdir.startswith(pkg_prefix):
+ libdir = libdir.replace(pkg_prefix, '${exec_prefix}')
+
+ includedir = bld.env['INCLUDEDIR']
+ if includedir.startswith(pkg_prefix):
+ includedir = includedir.replace(pkg_prefix, '${prefix}')
+
+ obj = bld(features='subst',
+ source='%s.pc.in' % name.lower(),
+ target=target,
+ install_path=os.path.join(bld.env['LIBDIR'], 'pkgconfig'),
+ exec_prefix='${prefix}',
+ PREFIX=pkg_prefix,
+ EXEC_PREFIX='${prefix}',
+ LIBDIR=libdir,
+ INCLUDEDIR=includedir)
+
+ if type(libs) != list:
+ libs = libs.split()
+
+ subst_dict[name + '_VERSION'] = version
+ subst_dict[name + '_MAJOR_VERSION'] = version[0:version.find('.')]
+ for i in libs:
+ subst_dict[i + '_LIBS'] = link_flags(bld.env, i)
+ lib_cflags = compile_flags(bld.env, i)
+ if lib_cflags == '':
+ lib_cflags = ' '
+ subst_dict[i + '_CFLAGS'] = lib_cflags
+
+ obj.__dict__.update(subst_dict)
+
+def build_dir(name, subdir):
+ if is_child():
+ return os.path.join('build', name, subdir)
+ else:
+ return os.path.join('build', subdir)
+
+
+def make_simple_dox(name):
+ "Clean up messy Doxygen documentation after it is built"
+ name = name.lower()
+ NAME = name.upper()
+ try:
+ top = os.getcwd()
+ os.chdir(build_dir(name, 'doc/html'))
+ page = 'group__%s.html' % name
+ if not os.path.exists(page):
+ return
+ for i in [
+ ['%s_API ' % NAME, ''],
+ ['%s_DEPRECATED ' % NAME, ''],
+ ['group__%s.html' % name, ''],
+ ['&#160;', ''],
+ [r'<script.*><\/script>', ''],
+ [r'<hr\/><a name="details" id="details"><\/a><h2>.*<\/h2>', ''],
+ [r'<link href=\"tabs.css\" rel=\"stylesheet\" type=\"text\/css\"\/>',
+ ''],
+ [r'<img class=\"footer\" src=\"doxygen.png\" alt=\"doxygen\"\/>',
+ 'Doxygen']]:
+ os.system("sed -i 's/%s/%s/g' %s" % (i[0], i[1], page))
+ os.rename('group__%s.html' % name, 'index.html')
+ for i in (glob.glob('*.png') +
+ glob.glob('*.html') +
+ glob.glob('*.js') +
+ glob.glob('*.css')):
+ if i != 'index.html' and i != 'style.css':
+ os.remove(i)
+ os.chdir(top)
+ os.chdir(build_dir(name, 'doc/man/man3'))
+ for i in glob.glob('*.3'):
+ os.system("sed -i 's/%s_API //' %s" % (NAME, i))
+ for i in glob.glob('_*'):
+ os.remove(i)
+ os.chdir(top)
+ except Exception as e:
+ Logs.error("Failed to fix up %s documentation: %s" % (name, e))
+ finally:
+ os.chdir(top)
+
+def build_dox(bld, name, version, srcdir, blddir, outdir='', versioned=True):
+ """Build Doxygen API documentation"""
+ if not bld.env['DOCS']:
+ return
+
+ # Doxygen paths in are relative to the doxygen file, not build directory
+ if is_child():
+ src_dir = os.path.join(srcdir, name.lower())
+ else:
+ src_dir = srcdir
+
+ subst_tg = bld(features='subst',
+ source='doc/reference.doxygen.in',
+ target='doc/reference.doxygen',
+ install_path='',
+ name='doxyfile')
+
+ subst_dict = {
+ name + '_VERSION': version,
+ name + '_SRCDIR': os.path.abspath(src_dir),
+ name + '_DOC_DIR': ''
+ }
+
+ subst_tg.__dict__.update(subst_dict)
+
+ subst_tg.post()
+
+ docs = bld(features='doxygen',
+ doxyfile='doc/reference.doxygen')
+
+ docs.post()
+
+ outname = name.lower()
+ if versioned:
+ outname += '-%d' % int(version[0:version.find('.')])
+ bld.install_files(
+ os.path.join('${DOCDIR}', outname, outdir, 'html'),
+ bld.path.get_bld().ant_glob('doc/html/*'))
+ for i in range(1, 8):
+ bld.install_files('${MANDIR}/man%d' % i,
+ bld.path.get_bld().ant_glob('doc/man/man%d/*' % i,
+ excl='**/_*'))
+
+
+def build_version_files(header_path, source_path, domain, major, minor, micro):
+ """Generate version code header"""
+ header_path = os.path.abspath(header_path)
+ source_path = os.path.abspath(source_path)
+ text = "int " + domain + "_major_version = " + str(major) + ";\n"
+ text += "int " + domain + "_minor_version = " + str(minor) + ";\n"
+ text += "int " + domain + "_micro_version = " + str(micro) + ";\n"
+ try:
+ o = open(source_path, 'w')
+ o.write(text)
+ o.close()
+ except IOError:
+ Logs.error('Failed to open %s for writing\n' % source_path)
+ sys.exit(-1)
+
+ text = "#ifndef __" + domain + "_version_h__\n"
+ text += "#define __" + domain + "_version_h__\n"
+ text += "extern const char* " + domain + "_revision;\n"
+ text += "extern int " + domain + "_major_version;\n"
+ text += "extern int " + domain + "_minor_version;\n"
+ text += "extern int " + domain + "_micro_version;\n"
+ text += "#endif /* __" + domain + "_version_h__ */\n"
+ try:
+ o = open(header_path, 'w')
+ o.write(text)
+ o.close()
+ except IOError:
+ Logs.warn('Failed to open %s for writing\n' % header_path)
+ sys.exit(-1)
+
+ return None
+
+def build_i18n_pot(bld, srcdir, dir, name, sources, copyright_holder=None):
+ Logs.info('Generating pot file from %s' % name)
+ pot_file = '%s.pot' % name
+
+ cmd = ['xgettext',
+ '--keyword=_',
+ '--keyword=N_',
+ '--keyword=S_',
+ '--from-code=UTF-8',
+ '-o', pot_file]
+
+ if copyright_holder:
+ cmd += ['--copyright-holder="%s"' % copyright_holder]
+
+ cmd += sources
+ Logs.info('Updating ' + pot_file)
+ subprocess.call(cmd, cwd=os.path.join(srcdir, dir))
+
+def build_i18n_po(bld, srcdir, dir, name, sources, copyright_holder=None):
+ pwd = os.getcwd()
+ os.chdir(os.path.join(srcdir, dir))
+ pot_file = '%s.pot' % name
+ po_files = glob.glob('po/*.po')
+ for po_file in po_files:
+ cmd = ['msgmerge',
+ '--update',
+ po_file,
+ pot_file]
+ Logs.info('Updating ' + po_file)
+ subprocess.call(cmd)
+ os.chdir(pwd)
+
+def build_i18n_mo(bld, srcdir, dir, name, sources, copyright_holder=None):
+ pwd = os.getcwd()
+ os.chdir(os.path.join(srcdir, dir))
+ po_files = glob.glob('po/*.po')
+ for po_file in po_files:
+ mo_file = po_file.replace('.po', '.mo')
+ cmd = ['msgfmt',
+ '-c',
+ '-f',
+ '-o',
+ mo_file,
+ po_file]
+ Logs.info('Generating ' + po_file)
+ subprocess.call(cmd)
+ os.chdir(pwd)
+
+def build_i18n(bld, srcdir, dir, name, sources, copyright_holder=None):
+ build_i18n_pot(bld, srcdir, dir, name, sources, copyright_holder)
+ build_i18n_po(bld, srcdir, dir, name, sources, copyright_holder)
+ build_i18n_mo(bld, srcdir, dir, name, sources, copyright_holder)
+
+class ExecutionEnvironment:
+ """Context that sets system environment variables for program execution"""
+ def __init__(self, changes):
+ self.original_environ = os.environ.copy()
+
+ self.diff = {}
+ for path_name, paths in changes.items():
+ value = os.pathsep.join(paths)
+ if path_name in os.environ:
+ value += os.pathsep + os.environ[path_name]
+
+ self.diff[path_name] = value
+
+ os.environ.update(self.diff)
+
+ def __str__(self):
+ return '\n'.join({'%s="%s"' % (k, v) for k, v in self.diff.items()})
+
+ def __enter__(self):
+ return self
+
+ def __exit__(self, type, value, traceback):
+ os.environ = self.original_environ
+
+class RunContext(Build.BuildContext):
+ "runs an executable from the build directory"
+ cmd = 'run'
+
+ def execute(self):
+ self.restore()
+ if not self.all_envs:
+ self.load_envs()
+
+ with ExecutionEnvironment(self.env.AUTOWAF_RUN_ENV) as env:
+ if Options.options.verbose:
+ Logs.pprint('GREEN', str(env) + '\n')
+
+ if Options.options.cmd:
+ Logs.pprint('GREEN', 'Running %s' % Options.options.cmd)
+ subprocess.call(Options.options.cmd, shell=True)
+ else:
+ Logs.error("error: Missing --cmd option for run command")
+
+def show_diff(from_lines, to_lines, from_filename, to_filename):
+ import difflib
+ import sys
+
+ same = True
+ for line in difflib.unified_diff(
+ from_lines, to_lines,
+ fromfile=os.path.abspath(from_filename),
+ tofile=os.path.abspath(to_filename)):
+ sys.stderr.write(line)
+ same = False
+
+ return same
+
+def test_file_equals(patha, pathb):
+ import filecmp
+ import io
+
+ for path in (patha, pathb):
+ if not os.access(path, os.F_OK):
+ Logs.pprint('RED', 'error: missing file %s' % path)
+ return False
+
+ if filecmp.cmp(patha, pathb, shallow=False):
+ return True
+
+ with io.open(patha, 'rU', encoding='utf-8') as fa:
+ with io.open(pathb, 'rU', encoding='utf-8') as fb:
+ return show_diff(fa.readlines(), fb.readlines(), patha, pathb)
+
+def bench_time():
+ if hasattr(time, 'perf_counter'): # Added in Python 3.3
+ return time.perf_counter()
+ else:
+ return time.time()
+
+class TestOutput:
+ """Test output that is truthy if result is as expected"""
+ def __init__(self, expected, result=None):
+ self.stdout = self.stderr = None
+ self.expected = expected
+ self.result = result
+
+ def __bool__(self):
+ return self.expected is None or self.result == self.expected
+
+ __nonzero__ = __bool__
+
+def is_string(s):
+ if sys.version_info[0] < 3:
+ return isinstance(s, basestring)
+ return isinstance(s, str)
+
+class TestScope:
+ """Scope for running tests that maintains pass/fail statistics"""
+ def __init__(self, tst, name, defaults):
+ self.tst = tst
+ self.name = name
+ self.defaults = defaults
+ self.n_failed = 0
+ self.n_total = 0
+
+ def run(self, test, **kwargs):
+ if type(test) == list and 'name' not in kwargs:
+ import pipes
+ kwargs['name'] = ' '.join(map(pipes.quote, test))
+
+ if Options.options.test_filter and 'name' in kwargs:
+ import re
+ found = False
+ for scope in self.tst.stack:
+ if re.search(Options.options.test_filter, scope.name):
+ found = True
+ break
+
+ if (not found and
+ not re.search(Options.options.test_filter, self.name) and
+ not re.search(Options.options.test_filter, kwargs['name'])):
+ return True
+
+ if callable(test):
+ output = self._run_callable(test, **kwargs)
+ elif type(test) == list:
+
+ output = self._run_command(test, **kwargs)
+ else:
+ raise Exception("Unknown test type")
+
+ if not output:
+ self.tst.log_bad('FAILED', kwargs['name'])
+
+ return self.tst.test_result(output)
+
+ def _run_callable(self, test, **kwargs):
+ expected = kwargs['expected'] if 'expected' in kwargs else True
+ return TestOutput(expected, test())
+
+ def _run_command(self, test, **kwargs):
+ if 'stderr' in kwargs and kwargs['stderr'] == NONEMPTY:
+ # Run with a temp file for stderr and check that it is non-empty
+ import tempfile
+ with tempfile.TemporaryFile() as stderr:
+ kwargs['stderr'] = stderr
+ output = self.run(test, **kwargs)
+ stderr.seek(0, 2) # Seek to end
+ return (output if not output else
+ self.run(
+ lambda: stderr.tell() > 0,
+ name=kwargs['name'] + ' error message'))
+
+ try:
+ # Run with stdout and stderr set to the appropriate streams
+ out_stream = self._stream('stdout', kwargs)
+ err_stream = self._stream('stderr', kwargs)
+ return self._exec(test, **kwargs)
+ finally:
+ out_stream = out_stream.close() if out_stream else None
+ err_stream = err_stream.close() if err_stream else None
+
+ def _stream(self, stream_name, kwargs):
+ s = kwargs[stream_name] if stream_name in kwargs else None
+ if is_string(s):
+ kwargs[stream_name] = open(s, 'wb')
+ return kwargs[stream_name]
+ return None
+
+ def _exec(self,
+ test,
+ expected=0,
+ name='',
+ stdin=None,
+ stdout=None,
+ stderr=None,
+ verbosity=1):
+ def stream(s):
+ return open(s, 'wb') if type(s) == str else s
+
+ if verbosity > 1:
+ self.tst.log_good('RUN ', name)
+
+ if Options.options.test_wrapper:
+ test = [Options.options.test_wrapper] + test
+
+ output = TestOutput(expected)
+ with open(os.devnull, 'wb') as null:
+ out = null if verbosity < 3 and not stdout else stdout
+ err = null if verbosity < 2 and not stderr else stderr
+ proc = subprocess.Popen(test, stdin=stdin, stdout=out, stderr=err)
+ output.stdout, output.stderr = proc.communicate()
+ output.result = proc.returncode
+
+ if output and verbosity > 0:
+ self.tst.log_good(' OK', name)
+
+ return output
+
+class TestContext(Build.BuildContext):
+ "runs test suite"
+ fun = cmd = 'test'
+
+ def __init__(self, **kwargs):
+ super(TestContext, self).__init__(**kwargs)
+ self.start_time = bench_time()
+ self.max_depth = 1
+
+ defaults = {'verbosity': Options.options.verbose}
+ self.stack = [TestScope(self, Context.g_module.APPNAME, defaults)]
+
+ def defaults(self):
+ return self.stack[-1].defaults
+
+ def finalize(self):
+ if self.stack[-1].n_failed > 0:
+ sys.exit(1)
+
+ super(TestContext, self).finalize()
+
+ def __call__(self, test, **kwargs):
+ return self.stack[-1].run(test, **self.args(**kwargs))
+
+ def file_equals(self, from_path, to_path, **kwargs):
+ kwargs.update({'expected': True,
+ 'name': '%s == %s' % (from_path, to_path)})
+ return self(lambda: test_file_equals(from_path, to_path), **kwargs)
+
+ def log_good(self, title, fmt, *args):
+ Logs.pprint('GREEN', '[%s] %s' % (title.center(10), fmt % args))
+
+ def log_bad(self, title, fmt, *args):
+ Logs.pprint('RED', '[%s] %s' % (title.center(10), fmt % args))
+
+ def pre_recurse(self, node):
+ wscript_module = Context.load_module(node.abspath())
+ group_name = wscript_module.APPNAME
+ self.stack.append(TestScope(self, group_name, self.defaults()))
+ self.max_depth = max(self.max_depth, len(self.stack) - 1)
+
+ bld_dir = node.get_bld().parent
+ if bld_dir != self.path.get_bld():
+ Logs.info('')
+
+ self.original_dir = os.getcwd()
+ Logs.info("Waf: Entering directory `%s'\n", bld_dir)
+ os.chdir(str(bld_dir))
+
+ if not self.env.NO_COVERAGE and str(node.parent) == Context.top_dir:
+ self.clear_coverage()
+
+ self.log_good('=' * 10, 'Running %s tests', group_name)
+ super(TestContext, self).pre_recurse(node)
+
+ def test_result(self, success):
+ self.stack[-1].n_total += 1
+ self.stack[-1].n_failed += 1 if not success else 0
+ return success
+
+ def pop(self):
+ scope = self.stack.pop()
+ self.stack[-1].n_total += scope.n_total
+ self.stack[-1].n_failed += scope.n_failed
+ return scope
+
+ def post_recurse(self, node):
+ super(TestContext, self).post_recurse(node)
+
+ scope = self.pop()
+ duration = (bench_time() - self.start_time) * 1000.0
+ is_top = str(node.parent) == str(Context.top_dir)
+
+ if is_top and self.max_depth > 1:
+ Logs.info('')
+
+ self.log_good('=' * 10, '%d tests from %s ran (%d ms total)',
+ scope.n_total, scope.name, duration)
+
+ if not self.env.NO_COVERAGE:
+ if is_top:
+ self.gen_coverage()
+
+ if os.path.exists('coverage/index.html'):
+ self.log_good('REPORT', '<file://%s>',
+ os.path.abspath('coverage/index.html'))
+
+ successes = scope.n_total - scope.n_failed
+ Logs.pprint('GREEN', '[ PASSED ] %d tests' % successes)
+ if scope.n_failed > 0:
+ Logs.pprint('RED', '[ FAILED ] %d tests' % scope.n_failed)
+ if is_top:
+ Logs.info("\nWaf: Leaving directory `%s'" % os.getcwd())
+
+ os.chdir(self.original_dir)
+
+ def execute(self):
+ self.restore()
+ if not self.all_envs:
+ self.load_envs()
+
+ if not self.env.BUILD_TESTS:
+ self.fatal('Configuration does not include tests')
+
+ with ExecutionEnvironment(self.env.AUTOWAF_RUN_ENV) as env:
+ if self.defaults()['verbosity'] > 0:
+ Logs.pprint('GREEN', str(env) + '\n')
+ self.recurse([self.run_dir])
+
+ def src_path(self, path):
+ return os.path.relpath(os.path.join(str(self.path), path))
+
+ def args(self, **kwargs):
+ all_kwargs = self.defaults().copy()
+ all_kwargs.update(kwargs)
+ return all_kwargs
+
+ def group(self, name, **kwargs):
+ return TestGroup(
+ self, self.stack[-1].name, name, **self.args(**kwargs))
+
+ def set_test_defaults(self, **kwargs):
+ """Set default arguments to be passed to all tests"""
+ self.stack[-1].defaults.update(kwargs)
+
+ def clear_coverage(self):
+ """Zero old coverage data"""
+ try:
+ with open('cov-clear.log', 'w') as log:
+ subprocess.call(['lcov', '-z', '-d', str(self.path)],
+ stdout=log, stderr=log)
+
+ except Exception:
+ Logs.warn('Failed to run lcov to clear old coverage data')
+
+ def gen_coverage(self):
+ """Generate coverage data and report"""
+ try:
+ with open('cov.lcov', 'w') as out:
+ with open('cov.log', 'w') as err:
+ subprocess.call(['lcov', '-c', '--no-external',
+ '--rc', 'lcov_branch_coverage=1',
+ '-b', '.',
+ '-d', str(self.path)],
+ stdout=out, stderr=err)
+
+ if not os.path.isdir('coverage'):
+ os.makedirs('coverage')
+
+ with open('genhtml.log', 'w') as log:
+ subprocess.call(['genhtml',
+ '-o', 'coverage',
+ '--rc', 'genhtml_branch_coverage=1',
+ 'cov.lcov'],
+ stdout=log, stderr=log)
+
+ summary = subprocess.check_output(
+ ['lcov', '--summary',
+ '--rc', 'lcov_branch_coverage=1',
+ 'cov.lcov'],
+ stderr=subprocess.STDOUT).decode('ascii')
+
+ import re
+ lines = re.search('lines\.*: (.*)%.*', summary).group(1)
+ functions = re.search('functions\.*: (.*)%.*', summary).group(1)
+ branches = re.search('branches\.*: (.*)%.*', summary).group(1)
+ self.log_good('COVERAGE', '%s%% lines, %s%% functions, %s%% branches',
+ lines, functions, branches)
+
+ except Exception:
+ Logs.warn('Failed to run lcov to generate coverage report')
+
+class TestGroup:
+ def __init__(self, tst, suitename, name, **kwargs):
+ self.tst = tst
+ self.suitename = suitename
+ self.name = name
+ self.kwargs = kwargs
+ self.start_time = bench_time()
+ tst.stack.append(TestScope(tst, name, tst.defaults()))
+
+ def label(self):
+ return self.suitename + '.%s' % self.name if self.name else ''
+
+ def args(self, **kwargs):
+ all_kwargs = self.tst.args(**self.kwargs)
+ all_kwargs.update(kwargs)
+ return all_kwargs
+
+ def __enter__(self):
+ if 'verbosity' in self.kwargs and self.kwargs['verbosity'] > 0:
+ self.tst.log_good('-' * 10, self.label())
+ return self
+
+ def __call__(self, test, **kwargs):
+ return self.tst(test, **self.args(**kwargs))
+
+ def file_equals(self, from_path, to_path, **kwargs):
+ return self.tst.file_equals(from_path, to_path, **kwargs)
+
+ def __exit__(self, type, value, traceback):
+ duration = (bench_time() - self.start_time) * 1000.0
+ scope = self.tst.pop()
+ n_passed = scope.n_total - scope.n_failed
+ if scope.n_failed == 0:
+ self.tst.log_good('-' * 10, '%d tests from %s (%d ms total)',
+ scope.n_total, self.label(), duration)
+ else:
+ self.tst.log_bad('-' * 10, '%d/%d tests from %s (%d ms total)',
+ n_passed, scope.n_total, self.label(), duration)
+
+def run_ldconfig(ctx):
+ should_run = (ctx.cmd == 'install' and
+ not ctx.env['RAN_LDCONFIG'] and
+ ctx.env['LIBDIR'] and
+ 'DESTDIR' not in os.environ and
+ not Options.options.destdir)
+
+ if should_run:
+ try:
+ Logs.info("Waf: Running `/sbin/ldconfig %s'" % ctx.env['LIBDIR'])
+ subprocess.call(['/sbin/ldconfig', ctx.env['LIBDIR']])
+ ctx.env['RAN_LDCONFIG'] = True
+ except Exception:
+ pass
+
+def get_rdf_news(name,
+ in_files,
+ top_entries=None,
+ extra_entries=None,
+ dev_dist=None):
+ import rdflib
+ from time import strptime
+
+ doap = rdflib.Namespace('http://usefulinc.com/ns/doap#')
+ dcs = rdflib.Namespace('http://ontologi.es/doap-changeset#')
+ rdfs = rdflib.Namespace('http://www.w3.org/2000/01/rdf-schema#')
+ foaf = rdflib.Namespace('http://xmlns.com/foaf/0.1/')
+ rdf = rdflib.Namespace('http://www.w3.org/1999/02/22-rdf-syntax-ns#')
+ m = rdflib.ConjunctiveGraph()
+
+ try:
+ for i in in_files:
+ m.parse(i, format='n3')
+ except Exception:
+ Logs.warn('Error parsing data, unable to generate NEWS')
+ return
+
+ proj = m.value(None, rdf.type, doap.Project)
+ for f in m.triples([proj, rdfs.seeAlso, None]):
+ if f[2].endswith('.ttl'):
+ m.parse(f[2], format='n3')
+
+ entries = {}
+ for r in m.triples([proj, doap.release, None]):
+ release = r[2]
+ revision = m.value(release, doap.revision, None)
+ date = m.value(release, doap.created, None)
+ blamee = m.value(release, dcs.blame, None)
+ changeset = m.value(release, dcs.changeset, None)
+ dist = m.value(release, doap['file-release'], None)
+
+ if not dist:
+ Logs.warn('No file release for %s %s' % (proj, revision))
+ dist = dev_dist
+
+ if revision and date and blamee and changeset:
+ entry = {}
+ entry['name'] = str(name)
+ entry['revision'] = str(revision)
+ entry['date'] = strptime(str(date), '%Y-%m-%d')
+ entry['status'] = 'stable' if dist != dev_dist else 'unstable'
+ entry['dist'] = str(dist)
+ entry['items'] = []
+
+ for i in m.triples([changeset, dcs.item, None]):
+ item = str(m.value(i[2], rdfs.label, None))
+ entry['items'] += [item]
+ if dist and top_entries is not None:
+ if not str(dist) in top_entries:
+ top_entries[str(dist)] = {'items': []}
+ top_entries[str(dist)]['items'] += [
+ '%s: %s' % (name, item)]
+
+ if extra_entries and dist:
+ for i in extra_entries[str(dist)]:
+ entry['items'] += extra_entries[str(dist)]['items']
+
+ entry['blamee_name'] = str(m.value(blamee, foaf.name, None))
+ entry['blamee_mbox'] = str(m.value(blamee, foaf.mbox, None))
+
+ entries[(str(date), str(revision))] = entry
+ else:
+ Logs.warn('Ignored incomplete %s release description' % name)
+
+ return entries
+
+def write_news(entries, out_file):
+ import textwrap
+ from time import strftime
+
+ if len(entries) == 0:
+ return
+
+ news = open(out_file, 'w')
+ for e in sorted(entries.keys(), reverse=True):
+ entry = entries[e]
+ news.write('%s (%s) %s;\n' % (entry['name'], entry['revision'], entry['status']))
+ for item in entry['items']:
+ wrapped = textwrap.wrap(item, width=79)
+ news.write('\n * ' + '\n '.join(wrapped))
+
+ news.write('\n\n --')
+ news.write(' %s <%s>' % (entry['blamee_name'],
+ entry['blamee_mbox'].replace('mailto:', '')))
+
+ news.write(' %s\n\n' % (
+ strftime('%a, %d %b %Y %H:%M:%S +0000', entry['date'])))
+
+ news.close()
+
+def write_posts(entries, meta, out_dir, status='stable'):
+ "write news posts in Pelican Markdown format"
+ from time import strftime
+ try:
+ os.mkdir(out_dir)
+ except Exception:
+ pass
+
+ for i in entries:
+ entry = entries[i]
+ revision = i[1]
+ if entry['status'] != status:
+ continue
+
+ date_str = strftime('%Y-%m-%d', entry['date'])
+ datetime_str = strftime('%Y-%m-%d %H:%M', entry['date'])
+
+ path = os.path.join(out_dir, '%s-%s-%s.md' % (
+ date_str, entry['name'], revision.replace('.', '-')))
+ post = open(path, 'w')
+ title = entry['title'] if 'title' in entry else entry['name']
+ post.write('Title: %s %s\n' % (title, revision))
+ post.write('Date: %s\n' % datetime_str)
+ post.write('Slug: %s-%s\n' % (entry['name'], revision.replace('.', '-')))
+ for k in meta:
+ post.write('%s: %s\n' % (k, meta[k]))
+ post.write('\n')
+
+ url = entry['dist']
+ if entry['status'] == status:
+ post.write('[%s %s](%s) has been released.' % (
+ (entry['name'], revision, url)))
+
+ if 'description' in entry:
+ post.write(' ' + entry['description'])
+
+ post.write('\n')
+ if (len(entry['items']) > 0 and
+ not (len(entry['items']) == 1 and
+ entry['items'][0] == 'Initial release')):
+ post.write('\nChanges:\n\n')
+ for i in entry['items']:
+ post.write(' * %s\n' % i)
+
+ post.close()
+
+def get_blurb(in_file):
+ "Get the first paragram of a Markdown formatted file, skipping the title"
+ f = open(in_file, 'r')
+ f.readline() # Title
+ f.readline() # Title underline
+ f.readline() # Blank
+ out = ''
+ line = f.readline()
+ while len(line) > 0 and line != '\n':
+ out += line.replace('\n', ' ')
+ line = f.readline()
+ return out.strip()
+
+def get_news(in_file, entry_props={}):
+ """Get NEWS entries in the format expected by write_posts().
+
+ Properties that should be set on every entry can be passed in
+ `entry_props`. If `entry_props` has a 'dist_pattern' value, it is used to
+ set the 'dist' entry of entries by substituting the version number.
+ """
+
+ import re
+ import rfc822
+
+ f = open(in_file, 'r')
+ entries = {}
+ while True:
+ # Read header line
+ head = f.readline()
+ matches = re.compile(r'([^ ]*) \((.*)\) ([a-zA-z]*);').match(head)
+ if matches is None:
+ break
+
+ entry = {}
+ entry['name'] = matches.group(1)
+ entry['revision'] = matches.group(2)
+ entry['status'] = matches.group(3)
+ entry['items'] = []
+ if 'dist_pattern' in entry_props:
+ entry['dist'] = entry_props['dist_pattern'] % entry['revision']
+
+ # Read blank line after header
+ if f.readline() != '\n':
+ raise SyntaxError('expected blank line after NEWS header')
+
+ def add_item(item):
+ if len(item) > 0:
+ entry['items'] += [item.replace('\n', ' ').strip()]
+
+ # Read entries for this revision
+ item = ''
+ line = ''
+ while line != '\n':
+ line = f.readline()
+ if line.startswith(' * '):
+ add_item(item)
+ item = line[3:].lstrip()
+ else:
+ item += line.lstrip()
+ add_item(item)
+
+ # Read footer line
+ foot = f.readline()
+ matches = re.compile(' -- (.*) <(.*)> (.*)').match(foot)
+ entry['date'] = rfc822.parsedate(matches.group(3))
+ entry['blamee_name'] = matches.group(1)
+ entry['blamee_mbox'] = matches.group(2)
+ entry.update(entry_props)
+ entries[(entry['date'], entry['revision'])] = entry
+
+ # Skip trailing blank line before next entry
+ f.readline()
+
+ f.close()
+
+ return entries
+
+def news_to_posts(news_file, entry_props, post_meta, default_post_dir):
+ post_dir = os.getenv('POST_DIR')
+ if not post_dir:
+ post_dir = default_post_dir
+ sys.stderr.write('POST_DIR not set in environment, writing to %s\n' % post_dir)
+ else:
+ sys.stderr.write('writing posts to %s\n' % post_dir)
+
+ entries = get_news(news_file, entry_props)
+ write_posts(entries, post_meta, post_dir)
+
+def run_script(cmds):
+ for cmd in cmds:
+ subprocess.check_call(cmd, shell=True)
+
+def release(name, version, dist_name=None):
+ if dist_name is None:
+ dist_name = name.lower()
+
+ dist = '%s-%s.tar.bz2' % (dist_name or name.lower(), version)
+ try:
+ os.remove(dist)
+ os.remove(dist + '.sig')
+ except Exception:
+ pass
+
+ status = subprocess.check_output('git status --porcelain', shell=True)
+ if status:
+ Logs.error('error: git working copy is dirty\n' + status)
+ raise Exception('git working copy is dirty')
+
+ head = subprocess.check_output('git show -s --oneline', shell=True)
+ head_summary = head[8:].strip().lower()
+ expected_summary = '%s %s' % (name.lower(), version)
+ if head_summary != expected_summary:
+ raise Exception('latest commit "%s" does not match "%s"' % (
+ head_summary, expected_summary))
+
+ run_script(['./waf configure --docs',
+ './waf',
+ './waf distcheck',
+ './waf posts',
+ 'gpg -b %s' % dist,
+ 'git tag -s v%s -m "%s %s"' % (version, name, version)])