158 lines
4.6 KiB
Python
158 lines
4.6 KiB
Python
|
#!/usr/bin/env python2
|
||
|
# SPDX-License-Identifier: GPL-2.0+
|
||
|
|
||
|
# Copyright (c) 2016 Google, Inc
|
||
|
# Written by Simon Glass <sjg@chromium.org>
|
||
|
#
|
||
|
# Creates binary images from input files controlled by a description
|
||
|
#
|
||
|
|
||
|
"""See README for more information"""
|
||
|
|
||
|
import glob
|
||
|
import multiprocessing
|
||
|
import os
|
||
|
import sys
|
||
|
import traceback
|
||
|
import unittest
|
||
|
|
||
|
# Bring in the patman and dtoc libraries
|
||
|
our_path = os.path.dirname(os.path.realpath(__file__))
|
||
|
for dirname in ['../patman', '../dtoc', '..', '../concurrencytest']:
|
||
|
sys.path.insert(0, os.path.join(our_path, dirname))
|
||
|
|
||
|
# Bring in the libfdt module
|
||
|
sys.path.insert(0, 'scripts/dtc/pylibfdt')
|
||
|
sys.path.insert(0, os.path.join(our_path,
|
||
|
'../../build-sandbox_spl/scripts/dtc/pylibfdt'))
|
||
|
|
||
|
import cmdline
|
||
|
import command
|
||
|
use_concurrent = True
|
||
|
try:
|
||
|
from concurrencytest import ConcurrentTestSuite, fork_for_tests
|
||
|
except:
|
||
|
use_concurrent = False
|
||
|
import control
|
||
|
import test_util
|
||
|
|
||
|
def RunTests(debug, processes, args):
|
||
|
"""Run the functional tests and any embedded doctests
|
||
|
|
||
|
Args:
|
||
|
debug: True to enable debugging, which shows a full stack trace on error
|
||
|
args: List of positional args provided to binman. This can hold a test
|
||
|
name to execute (as in 'binman -t testSections', for example)
|
||
|
processes: Number of processes to use to run tests (None=same as #CPUs)
|
||
|
"""
|
||
|
import elf_test
|
||
|
import entry_test
|
||
|
import fdt_test
|
||
|
import ftest
|
||
|
import image_test
|
||
|
import test
|
||
|
import doctest
|
||
|
|
||
|
result = unittest.TestResult()
|
||
|
for module in []:
|
||
|
suite = doctest.DocTestSuite(module)
|
||
|
suite.run(result)
|
||
|
|
||
|
sys.argv = [sys.argv[0]]
|
||
|
if debug:
|
||
|
sys.argv.append('-D')
|
||
|
if debug:
|
||
|
sys.argv.append('-D')
|
||
|
|
||
|
# Run the entry tests first ,since these need to be the first to import the
|
||
|
# 'entry' module.
|
||
|
test_name = args and args[0] or None
|
||
|
suite = unittest.TestSuite()
|
||
|
loader = unittest.TestLoader()
|
||
|
for module in (entry_test.TestEntry, ftest.TestFunctional, fdt_test.TestFdt,
|
||
|
elf_test.TestElf, image_test.TestImage):
|
||
|
if test_name:
|
||
|
try:
|
||
|
suite.addTests(loader.loadTestsFromName(test_name, module))
|
||
|
except AttributeError:
|
||
|
continue
|
||
|
else:
|
||
|
suite.addTests(loader.loadTestsFromTestCase(module))
|
||
|
if use_concurrent and processes != 1:
|
||
|
concurrent_suite = ConcurrentTestSuite(suite,
|
||
|
fork_for_tests(processes or multiprocessing.cpu_count()))
|
||
|
concurrent_suite.run(result)
|
||
|
else:
|
||
|
suite.run(result)
|
||
|
|
||
|
print result
|
||
|
for test, err in result.errors:
|
||
|
print test.id(), err
|
||
|
for test, err in result.failures:
|
||
|
print err, result.failures
|
||
|
if result.errors or result.failures:
|
||
|
print 'binman tests FAILED'
|
||
|
return 1
|
||
|
return 0
|
||
|
|
||
|
def GetEntryModules(include_testing=True):
|
||
|
"""Get a set of entry class implementations
|
||
|
|
||
|
Returns:
|
||
|
Set of paths to entry class filenames
|
||
|
"""
|
||
|
glob_list = glob.glob(os.path.join(our_path, 'etype/*.py'))
|
||
|
return set([os.path.splitext(os.path.basename(item))[0]
|
||
|
for item in glob_list
|
||
|
if include_testing or '_testing' not in item])
|
||
|
|
||
|
def RunTestCoverage():
|
||
|
"""Run the tests and check that we get 100% coverage"""
|
||
|
glob_list = GetEntryModules(False)
|
||
|
all_set = set([os.path.splitext(os.path.basename(item))[0]
|
||
|
for item in glob_list if '_testing' not in item])
|
||
|
test_util.RunTestCoverage('tools/binman/binman.py', None,
|
||
|
['*test*', '*binman.py', 'tools/patman/*', 'tools/dtoc/*'],
|
||
|
options.build_dir, all_set)
|
||
|
|
||
|
def RunBinman(options, args):
|
||
|
"""Main entry point to binman once arguments are parsed
|
||
|
|
||
|
Args:
|
||
|
options: Command-line options
|
||
|
args: Non-option arguments
|
||
|
"""
|
||
|
ret_code = 0
|
||
|
|
||
|
# For testing: This enables full exception traces.
|
||
|
#options.debug = True
|
||
|
|
||
|
if not options.debug:
|
||
|
sys.tracebacklimit = 0
|
||
|
|
||
|
if options.test:
|
||
|
ret_code = RunTests(options.debug, options.processes, args[1:])
|
||
|
|
||
|
elif options.test_coverage:
|
||
|
RunTestCoverage()
|
||
|
|
||
|
elif options.entry_docs:
|
||
|
control.WriteEntryDocs(GetEntryModules())
|
||
|
|
||
|
else:
|
||
|
try:
|
||
|
ret_code = control.Binman(options, args)
|
||
|
except Exception as e:
|
||
|
print 'binman: %s' % e
|
||
|
if options.debug:
|
||
|
print
|
||
|
traceback.print_exc()
|
||
|
ret_code = 1
|
||
|
return ret_code
|
||
|
|
||
|
|
||
|
if __name__ == "__main__":
|
||
|
(options, args) = cmdline.ParseArgs(sys.argv)
|
||
|
ret_code = RunBinman(options, args)
|
||
|
sys.exit(ret_code)
|