mirror of
https://github.com/dart-lang/sdk
synced 2024-11-02 08:20:31 +00:00
705be03363
BUG= TEST= Review URL: https://chromereviews.googleplex.com/3521020 git-svn-id: https://dart.googlecode.com/svn/branches/bleeding_edge/dart@52 260f80e4-7a28-3924-810f-c04153c831b5
1503 lines
42 KiB
Python
Executable file
1503 lines
42 KiB
Python
Executable file
#!/usr/bin/env python
|
|
#
|
|
# Copyright (c) 2011, the Dart project authors. Please see the AUTHORS file
|
|
# for details. All rights reserved. Use of this source code is governed by a
|
|
# BSD-style license that can be found in the LICENSE file.
|
|
#
|
|
|
|
import imp
|
|
import optparse
|
|
import os
|
|
from os.path import join, dirname, abspath, basename, isdir, exists, realpath
|
|
import platform
|
|
import re
|
|
import run
|
|
import select
|
|
import signal
|
|
import subprocess
|
|
import sys
|
|
import tempfile
|
|
import time
|
|
import threading
|
|
import traceback
|
|
from Queue import Queue, Empty
|
|
|
|
import testing
|
|
import testing.test_runner
|
|
import utils
|
|
|
|
|
|
TIMEOUT_SECS = 60
|
|
VERBOSE = False
|
|
ARCH_GUESS = utils.GuessArchitecture()
|
|
OS_GUESS = utils.GuessOS()
|
|
BUILT_IN_TESTS = ['dartc', 'vm', 'dart', 'corelib', 'language', 'co19',
|
|
'samples', 'isolate', 'stub-generator', 'client']
|
|
|
|
# Patterns for matching test options in .dart files.
|
|
VM_OPTIONS_PATTERN = re.compile(r"// VMOptions=(.*)")
|
|
DART_OPTIONS_PATTERN = re.compile(r"// DartOptions=(.*)")
|
|
ISOLATE_STUB_PATTERN = re.compile(r"// IsolateStubs=(.*)")
|
|
|
|
# ---------------------------------------------
|
|
# --- P r o g r e s s I n d i c a t o r s ---
|
|
# ---------------------------------------------
|
|
|
|
|
|
class ProgressIndicator(object):
|
|
|
|
def __init__(self, cases, context):
|
|
self.abort = False
|
|
self.terminate = False
|
|
self.cases = cases
|
|
self.queue = Queue(len(cases))
|
|
self.batch_queues = {};
|
|
self.context = context
|
|
|
|
# Extract batchable cases.
|
|
found_cmds = {}
|
|
batch_cases = []
|
|
for case in cases:
|
|
cmd = case.case.GetCommand()[0]
|
|
if not utils.IsWindows():
|
|
# Diagnostic check for executable (if an absolute pathname)
|
|
if not cmd in found_cmds:
|
|
if os.path.isabs(cmd) and not os.path.isfile(cmd):
|
|
msg = "Can't find command %s\n" % cmd \
|
|
+ "(Did you build first? " \
|
|
+ "Are you running in the correct directory?)"
|
|
raise Exception(msg)
|
|
else:
|
|
found_cmds[cmd] = 1
|
|
|
|
if case.case.IsBatchable():
|
|
if not self.batch_queues.has_key(cmd):
|
|
self.batch_queues[cmd] = Queue(len(cases))
|
|
self.batch_queues[cmd].put(case)
|
|
else:
|
|
self.queue.put_nowait(case)
|
|
|
|
self.succeeded = 0
|
|
self.remaining = len(cases)
|
|
self.total = len(cases)
|
|
self.failed = [ ]
|
|
self.crashed = 0
|
|
self.lock = threading.Lock()
|
|
|
|
def PrintFailureHeader(self, test):
|
|
if test.IsNegative():
|
|
negative_marker = '[negative] '
|
|
else:
|
|
negative_marker = ''
|
|
print "=== %(label)s %(negative)s===" % {
|
|
'label': test.GetLabel(),
|
|
'negative': negative_marker
|
|
}
|
|
print "Path: %s" % "/".join(test.path)
|
|
|
|
def Run(self, tasks):
|
|
self.Starting()
|
|
|
|
# Scale the number of tasks to the nubmer of CPUs on the machine
|
|
if tasks == testing.USE_DEFAULT_CPUS:
|
|
tasks = testing.HOST_CPUS
|
|
|
|
# TODO(zundel): Refactor BatchSingle method and TestRunner to
|
|
# share code and simplify this method.
|
|
|
|
# Start the non-batchable items first - there are some long running
|
|
# jobs we don't want to wait on at the end.
|
|
threads = []
|
|
# Spawn N-1 threads and then use this thread as the last one.
|
|
# That way -j1 avoids threading altogether which is a nice fallback
|
|
# in case of threading problems.
|
|
for i in xrange(tasks - 1):
|
|
thread = threading.Thread(target=self.RunSingle, args=[])
|
|
threads.append(thread)
|
|
thread.start()
|
|
|
|
|
|
# Next, crank up the batchable tasks. Note that this will start
|
|
# 'tasks' more threads, but the assumption is that if batching is
|
|
# enabled that almost all tests are batchable.
|
|
for (cmd, queue) in self.batch_queues.items():
|
|
if not queue.empty():
|
|
batch_len = queue.qsize();
|
|
batch_tester = None
|
|
try:
|
|
batch_tester = testing.test_runner.BatchRunner(queue, tasks, self,
|
|
[cmd, '-batch'])
|
|
except Exception, e:
|
|
print "Aborting batch test for " + cmd + ". Problem on startup."
|
|
if batch_tester: batch_tester.Shutdown()
|
|
raise
|
|
|
|
try:
|
|
batch_tester.WaitForCompletion()
|
|
except:
|
|
print "Aborting batch cmd " + cmd + "while waiting for completion."
|
|
if batch_tester: batch_tester.Shutdown()
|
|
raise
|
|
|
|
try:
|
|
self.RunSingle()
|
|
if self.abort:
|
|
raise Exception("Aborted")
|
|
# Wait for the remaining non-batched threads.
|
|
for thread in threads:
|
|
# Use a timeout so that signals (ctrl-c) will be processed.
|
|
thread.join(timeout=10000000)
|
|
if self.abort:
|
|
raise Exception("Aborted")
|
|
except Exception, e:
|
|
# If there's an exception we schedule an interruption for any
|
|
# remaining threads.
|
|
self.terminate = True
|
|
# ...and then reraise the exception to bail out
|
|
raise
|
|
|
|
self.Done()
|
|
return not self.failed
|
|
|
|
def RunSingle(self):
|
|
while not self.terminate:
|
|
try:
|
|
test = self.queue.get_nowait()
|
|
except Empty:
|
|
return
|
|
case = test.case
|
|
with self.lock:
|
|
self.AboutToRun(case)
|
|
try:
|
|
start = time.time()
|
|
output = case.Run()
|
|
case.duration = (time.time() - start)
|
|
except KeyboardInterrupt:
|
|
self.abort = True
|
|
self.terminate = True
|
|
raise
|
|
except IOError, e:
|
|
self.abort = True
|
|
self.terminate = True
|
|
raise
|
|
if self.terminate:
|
|
return
|
|
with self.lock:
|
|
if output.UnexpectedOutput():
|
|
self.failed.append(output)
|
|
if output.HasCrashed():
|
|
self.crashed += 1
|
|
else:
|
|
self.succeeded += 1
|
|
self.remaining -= 1
|
|
self.HasRun(output)
|
|
|
|
|
|
def EscapeCommand(command):
|
|
parts = []
|
|
for part in command:
|
|
if ' ' in part:
|
|
# Escape spaces. We may need to escape more characters for this
|
|
# to work properly.
|
|
parts.append('"%s"' % part)
|
|
else:
|
|
parts.append(part)
|
|
return " ".join(parts)
|
|
|
|
|
|
class SimpleProgressIndicator(ProgressIndicator):
|
|
|
|
def Starting(self):
|
|
print 'Running %i tests' % len(self.cases)
|
|
|
|
def Done(self):
|
|
print
|
|
for failed in self.failed:
|
|
self.PrintFailureHeader(failed.test)
|
|
if failed.output.stderr:
|
|
print "--- stderr ---"
|
|
print failed.output.stderr.strip()
|
|
if failed.output.stdout:
|
|
print "--- stdout ---"
|
|
print failed.output.stdout.strip()
|
|
print "Command: %s" % EscapeCommand(failed.command)
|
|
if failed.HasCrashed():
|
|
print "--- CRASHED ---"
|
|
if failed.HasTimedOut():
|
|
print "--- TIMEOUT ---"
|
|
if len(self.failed) == 0:
|
|
print "==="
|
|
print "=== All tests succeeded"
|
|
print "==="
|
|
else:
|
|
print
|
|
print "==="
|
|
if len(self.failed) == 1:
|
|
print "=== 1 test failed"
|
|
else:
|
|
print "=== %i tests failed" % len(self.failed)
|
|
if self.crashed > 0:
|
|
if self.crashed == 1:
|
|
print "=== 1 test CRASHED"
|
|
else:
|
|
print "=== %i tests CRASHED" % self.crashed
|
|
print "==="
|
|
|
|
|
|
class VerboseProgressIndicator(SimpleProgressIndicator):
|
|
|
|
def AboutToRun(self, case):
|
|
print 'Starting %s...' % case.GetLabel()
|
|
sys.stdout.flush()
|
|
|
|
def HasRun(self, output):
|
|
if output.UnexpectedOutput():
|
|
if output.HasCrashed():
|
|
outcome = 'CRASH'
|
|
else:
|
|
outcome = 'FAIL'
|
|
else:
|
|
outcome = 'PASS'
|
|
print 'Done running %s: %s' % (output.test.GetLabel(), outcome)
|
|
|
|
|
|
class OneLineProgressIndicator(SimpleProgressIndicator):
|
|
|
|
def AboutToRun(self, case):
|
|
pass
|
|
|
|
def HasRun(self, output):
|
|
if output.UnexpectedOutput():
|
|
if output.HasCrashed():
|
|
outcome = 'CRASH'
|
|
else:
|
|
outcome = 'FAIL'
|
|
else:
|
|
outcome = 'pass'
|
|
print 'Done %s: %s' % (output.test.GetLabel(), outcome)
|
|
|
|
|
|
class OneLineProgressIndicatorForBuildBot(OneLineProgressIndicator):
|
|
|
|
def HasRun(self, output):
|
|
super(OneLineProgressIndicatorForBuildBot, self).HasRun(output)
|
|
percent = (((self.total - self.remaining) * 100) // self.total)
|
|
print '@@@STEP_CLEAR@@@'
|
|
print '@@@STEP_TEXT@ %3d%% +%d -%d @@@' % (
|
|
percent, self.succeeded, len(self.failed))
|
|
|
|
|
|
class CompactProgressIndicator(ProgressIndicator):
|
|
|
|
def __init__(self, cases, context, templates):
|
|
super(CompactProgressIndicator, self).__init__(cases, context)
|
|
self.templates = templates
|
|
self.last_status_length = 0
|
|
self.start_time = time.time()
|
|
|
|
def Starting(self):
|
|
pass
|
|
|
|
def Done(self):
|
|
self.PrintProgress('Done')
|
|
|
|
def AboutToRun(self, case):
|
|
self.PrintProgress(case.GetLabel())
|
|
|
|
def HasRun(self, output):
|
|
if output.UnexpectedOutput():
|
|
self.ClearLine(self.last_status_length)
|
|
self.PrintFailureHeader(output.test)
|
|
stdout = output.output.stdout.strip()
|
|
if len(stdout):
|
|
print self.templates['stdout'] % stdout
|
|
stderr = output.output.stderr.strip()
|
|
if len(stderr):
|
|
print self.templates['stderr'] % stderr
|
|
print "Command: %s" % EscapeCommand(output.command)
|
|
if output.HasCrashed():
|
|
print "--- CRASHED ---"
|
|
if output.HasTimedOut():
|
|
print "--- TIMEOUT ---"
|
|
|
|
def Truncate(self, str, length):
|
|
if length and (len(str) > (length - 3)):
|
|
return str[:(length-3)] + "..."
|
|
else:
|
|
return str
|
|
|
|
def PrintProgress(self, name):
|
|
self.ClearLine(self.last_status_length)
|
|
elapsed = time.time() - self.start_time
|
|
status = self.templates['status_line'] % {
|
|
'passed': self.succeeded,
|
|
'percent': (((self.total - self.remaining) * 100) // self.total),
|
|
'failed': len(self.failed),
|
|
'test': name,
|
|
'mins': int(elapsed) / 60,
|
|
'secs': int(elapsed) % 60
|
|
}
|
|
status = self.Truncate(status, 78)
|
|
self.last_status_length = len(status)
|
|
print status,
|
|
sys.stdout.flush()
|
|
|
|
|
|
class MonochromeProgressIndicator(CompactProgressIndicator):
|
|
|
|
def __init__(self, cases, context):
|
|
templates = {
|
|
'status_line': "[%(mins)02i:%(secs)02i|%%%(percent) 4d|+%(passed) 4d|-%(failed) 4d]: %(test)s",
|
|
'stdout': '%s',
|
|
'stderr': '%s',
|
|
'clear': lambda last_line_length: ("\r" + (" " * last_line_length) + "\r"),
|
|
'max_length': 78
|
|
}
|
|
super(MonochromeProgressIndicator, self).__init__(cases, context, templates)
|
|
|
|
def ClearLine(self, last_line_length):
|
|
print ("\r" + (" " * last_line_length) + "\r"),
|
|
|
|
class ColorProgressIndicator(CompactProgressIndicator):
|
|
|
|
def __init__(self, cases, context):
|
|
templates = {
|
|
'status_line': ("[%(mins)02i:%(secs)02i|%%%(percent) 4d|"
|
|
"\033[32m+%(passed) 4d"
|
|
"\033[0m|\033[31m-%(failed) 4d\033[0m]: %(test)s"),
|
|
'stdout': '%s',
|
|
'stderr': '%s',
|
|
'clear': lambda last_line_length: ("\r" + (" " * last_line_length) + "\r"),
|
|
'max_length': 78
|
|
}
|
|
super(ColorProgressIndicator, self).__init__(cases, context, templates)
|
|
|
|
def ClearLine(self, last_line_length):
|
|
print ("\r" + (" " * last_line_length) + "\r"),
|
|
|
|
|
|
PROGRESS_INDICATORS = {
|
|
'verbose': VerboseProgressIndicator,
|
|
'mono': MonochromeProgressIndicator,
|
|
'color': ColorProgressIndicator,
|
|
'line': OneLineProgressIndicator,
|
|
'buildbot': OneLineProgressIndicatorForBuildBot
|
|
}
|
|
|
|
|
|
# -------------------------
|
|
# --- F r a m e w o r k ---
|
|
# -------------------------
|
|
|
|
|
|
class CommandOutput(object):
|
|
|
|
def __init__(self, pid, exit_code, timed_out, stdout, stderr):
|
|
self.pid = pid
|
|
self.exit_code = exit_code
|
|
self.timed_out = timed_out
|
|
self.stdout = stdout
|
|
self.stderr = stderr
|
|
self.failed = None
|
|
|
|
|
|
class TestCase(object):
|
|
|
|
def __init__(self, context, path):
|
|
self.path = path
|
|
self.context = context
|
|
self.duration = None
|
|
self.arch = []
|
|
|
|
def IsBatchable(self):
|
|
if self.context.use_batch:
|
|
if self.arch and 'dartc' in self.arch:
|
|
return True
|
|
return False
|
|
|
|
def IsNegative(self):
|
|
return False
|
|
|
|
def CompareTime(self, other):
|
|
return cmp(other.duration, self.duration)
|
|
|
|
def DidFail(self, output):
|
|
if output.failed is None:
|
|
output.failed = self.IsFailureOutput(output)
|
|
return output.failed
|
|
|
|
def IsFailureOutput(self, output):
|
|
return output.exit_code != 0
|
|
|
|
def RunCommand(self, command, cwd=None):
|
|
full_command = self.context.processor(command)
|
|
try:
|
|
output = Execute(full_command, self.context, self.context.timeout, cwd)
|
|
except OSError as e:
|
|
raise utils.ToolError("%s: %s" % (full_command[0], e.strerror))
|
|
test_output = TestOutput(self, full_command, output)
|
|
self.Cleanup()
|
|
return test_output
|
|
|
|
def BeforeRun(self):
|
|
pass
|
|
|
|
def AfterRun(self):
|
|
pass
|
|
|
|
def Run(self):
|
|
self.BeforeRun()
|
|
cmd = self.GetCommand()
|
|
try:
|
|
result = self.RunCommand(cmd)
|
|
finally:
|
|
self.AfterRun()
|
|
return result
|
|
|
|
def Cleanup(self):
|
|
return
|
|
|
|
|
|
class TestOutput(object):
|
|
|
|
def __init__(self, test, command, output):
|
|
self.test = test
|
|
self.command = command
|
|
self.output = output
|
|
|
|
def UnexpectedOutput(self):
|
|
if self.HasCrashed():
|
|
outcome = testing.CRASH
|
|
elif self.HasTimedOut():
|
|
outcome = TIMEOUT
|
|
elif self.HasFailed():
|
|
outcome = testing.FAIL
|
|
else:
|
|
outcome = testing.PASS
|
|
return not outcome in self.test.outcomes
|
|
|
|
def HasCrashed(self):
|
|
if utils.IsWindows():
|
|
if self.output.exit_code == 3:
|
|
# The VM uses std::abort to terminate on asserts.
|
|
# std::abort terminates with exit code 3 on Windows.
|
|
return True
|
|
return 0x80000000 & self.output.exit_code and not (0x3FFFFF00 & self.output.exit_code)
|
|
else:
|
|
# Timed out tests will have exit_code -signal.SIGTERM.
|
|
if self.output.timed_out:
|
|
return False
|
|
if self.output.exit_code == 253:
|
|
# The Java dartc runners exit 253 in case of unhandled exceptions.
|
|
return True
|
|
return self.output.exit_code < 0
|
|
|
|
def HasTimedOut(self):
|
|
return self.output.timed_out;
|
|
|
|
def HasFailed(self):
|
|
execution_failed = self.test.DidFail(self.output)
|
|
if self.test.IsNegative():
|
|
return not execution_failed
|
|
else:
|
|
return execution_failed
|
|
|
|
|
|
def KillProcessWithID(pid):
|
|
if utils.IsWindows():
|
|
os.popen('taskkill /T /F /PID %d' % pid)
|
|
else:
|
|
os.kill(pid, signal.SIGTERM)
|
|
|
|
|
|
MAX_SLEEP_TIME = 0.1
|
|
INITIAL_SLEEP_TIME = 0.0001
|
|
SLEEP_TIME_FACTOR = 1.25
|
|
|
|
SEM_INVALID_VALUE = -1
|
|
SEM_NOGPFAULTERRORBOX = 0x0002 # Microsoft Platform SDK WinBase.h
|
|
|
|
def Win32SetErrorMode(mode):
|
|
prev_error_mode = SEM_INVALID_VALUE
|
|
try:
|
|
import ctypes
|
|
prev_error_mode = ctypes.windll.kernel32.SetErrorMode(mode);
|
|
except ImportError:
|
|
pass
|
|
return prev_error_mode
|
|
|
|
def RunProcess(context, timeout, args, **rest):
|
|
if context.verbose: print "#", " ".join(args)
|
|
popen_args = args
|
|
prev_error_mode = SEM_INVALID_VALUE;
|
|
if utils.IsWindows():
|
|
popen_args = '"' + subprocess.list2cmdline(args) + '"'
|
|
if context.suppress_dialogs:
|
|
# Try to change the error mode to avoid dialogs on fatal errors. Don't
|
|
# touch any existing error mode flags by merging the existing error mode.
|
|
# See http://blogs.msdn.com/oldnewthing/archive/2004/07/27/198410.aspx.
|
|
error_mode = SEM_NOGPFAULTERRORBOX;
|
|
prev_error_mode = Win32SetErrorMode(error_mode);
|
|
Win32SetErrorMode(error_mode | prev_error_mode);
|
|
process = subprocess.Popen(
|
|
shell = utils.IsWindows(),
|
|
args = popen_args,
|
|
**rest
|
|
)
|
|
if utils.IsWindows() and context.suppress_dialogs and prev_error_mode != SEM_INVALID_VALUE:
|
|
Win32SetErrorMode(prev_error_mode)
|
|
# Compute the end time - if the process crosses this limit we
|
|
# consider it timed out.
|
|
if timeout is None: end_time = None
|
|
else: end_time = time.time() + timeout
|
|
timed_out = False
|
|
# Repeatedly check the exit code from the process in a
|
|
# loop and keep track of whether or not it times out.
|
|
exit_code = None
|
|
sleep_time = INITIAL_SLEEP_TIME
|
|
while exit_code is None:
|
|
if (not end_time is None) and (time.time() >= end_time):
|
|
# Kill the process and wait for it to exit.
|
|
KillProcessWithID(process.pid)
|
|
# Drain the output pipe from the process to avoid deadlock
|
|
process.communicate()
|
|
exit_code = process.wait()
|
|
timed_out = True
|
|
else:
|
|
exit_code = process.poll()
|
|
time.sleep(sleep_time)
|
|
sleep_time = sleep_time * SLEEP_TIME_FACTOR
|
|
if sleep_time > MAX_SLEEP_TIME:
|
|
sleep_time = MAX_SLEEP_TIME
|
|
return (process, exit_code, timed_out)
|
|
|
|
|
|
def PrintError(str):
|
|
sys.stderr.write(str)
|
|
sys.stderr.write('\n')
|
|
|
|
|
|
def CheckedUnlink(name):
|
|
try:
|
|
os.unlink(name)
|
|
except OSError, e:
|
|
PrintError("os.unlink() " + str(e))
|
|
|
|
|
|
def Execute(args, context, timeout=None, cwd=None):
|
|
(fd_out, outname) = tempfile.mkstemp()
|
|
(fd_err, errname) = tempfile.mkstemp()
|
|
(process, exit_code, timed_out) = RunProcess(
|
|
context,
|
|
timeout,
|
|
args = args,
|
|
stdout = fd_out,
|
|
stderr = fd_err,
|
|
cwd = cwd
|
|
)
|
|
os.close(fd_out)
|
|
os.close(fd_err)
|
|
output = file(outname).read()
|
|
errors = file(errname).read()
|
|
CheckedUnlink(outname)
|
|
CheckedUnlink(errname)
|
|
result = CommandOutput(process.pid, exit_code, timed_out, output, errors)
|
|
return result
|
|
|
|
|
|
class TestConfiguration(object):
|
|
|
|
def __init__(self, context, root):
|
|
self.context = context
|
|
self.root = root
|
|
|
|
def Contains(self, path, file):
|
|
if len(path) > len(file):
|
|
return False
|
|
for i in xrange(len(path)):
|
|
if not path[i].match(file[i]):
|
|
return False
|
|
return True
|
|
|
|
def GetTestStatus(self, sections, defs):
|
|
pass
|
|
|
|
|
|
class TestSuite(object):
|
|
|
|
def __init__(self, name):
|
|
self.name = name
|
|
|
|
def GetName(self):
|
|
return self.name
|
|
|
|
|
|
class TestRepository(TestSuite):
|
|
|
|
def __init__(self, path):
|
|
normalized_path = abspath(path)
|
|
super(TestRepository, self).__init__(basename(normalized_path))
|
|
self.path = normalized_path
|
|
self.is_loaded = False
|
|
self.config = None
|
|
|
|
def GetConfiguration(self, context):
|
|
if self.is_loaded:
|
|
return self.config
|
|
self.is_loaded = True
|
|
file = None
|
|
try:
|
|
(file, pathname, description) = imp.find_module('testcfg', [ self.path ])
|
|
module = imp.load_module('testcfg', file, pathname, description)
|
|
self.config = module.GetConfiguration(context, self.path)
|
|
finally:
|
|
if file:
|
|
file.close()
|
|
return self.config
|
|
|
|
def ListTests(self, current_path, path, context, mode, arch):
|
|
return self.GetConfiguration(context).ListTests(current_path,
|
|
path,
|
|
mode,
|
|
arch)
|
|
|
|
def GetTestStatus(self, context, sections, defs):
|
|
self.GetConfiguration(context).GetTestStatus(sections, defs)
|
|
|
|
|
|
class LiteralTestSuite(TestSuite):
|
|
|
|
def __init__(self, tests):
|
|
super(LiteralTestSuite, self).__init__('root')
|
|
self.tests = tests
|
|
|
|
def ListTests(self, current_path, path, context, mode, arch):
|
|
name = path[0]
|
|
result = [ ]
|
|
for test in self.tests:
|
|
test_name = test.GetName()
|
|
if name.match(test_name):
|
|
full_path = current_path + [test_name]
|
|
result += test.ListTests(full_path, path, context, mode, arch)
|
|
return result
|
|
|
|
def GetTestStatus(self, context, sections, defs):
|
|
for test in self.tests:
|
|
test.GetTestStatus(context, sections, defs)
|
|
|
|
class Context(object):
|
|
|
|
def __init__(self, workspace, verbose, os, timeout,
|
|
processor, suppress_dialogs, executable, flags,
|
|
keep_temporary_files, use_batch):
|
|
self.workspace = workspace
|
|
self.verbose = verbose
|
|
self.os = os
|
|
self.timeout = timeout
|
|
self.processor = processor
|
|
self.suppress_dialogs = suppress_dialogs
|
|
self.executable = executable
|
|
self.flags = flags
|
|
self.keep_temporary_files = keep_temporary_files
|
|
self.use_batch = use_batch == "true"
|
|
|
|
def GetBuildRoot(self, mode, arch):
|
|
result = utils.GetBuildRoot(self.os, mode, arch)
|
|
return result
|
|
|
|
def GetBuildConf(self, mode, arch):
|
|
result = utils.GetBuildConf(mode, arch)
|
|
return result
|
|
|
|
def GetExecutable(self, mode, arch, name):
|
|
if self.executable is not None:
|
|
return self.executable
|
|
path = abspath(join(self.GetBuildRoot(mode, arch), name))
|
|
if utils.IsWindows() and not path.endswith('.exe'):
|
|
return path + '.exe'
|
|
else:
|
|
return path
|
|
|
|
def GetDart(self, mode, arch):
|
|
if arch == 'dartc':
|
|
command = [ abspath(join(self.GetBuildRoot(mode, arch),
|
|
'compiler', 'bin', 'dartc_test')) ]
|
|
else:
|
|
command = [ self.GetExecutable(mode, arch, 'dart_bin') ]
|
|
|
|
return command
|
|
|
|
def GetDartC(self, mode, arch):
|
|
dartc = abspath(os.path.join(self.GetBuildRoot(mode, arch),
|
|
'compiler', 'bin', 'dartc'))
|
|
if utils.IsWindows(): dartc += '.exe'
|
|
command = [ dartc ]
|
|
|
|
# Add the flags from the context to the command line.
|
|
command += self.flags
|
|
return command
|
|
|
|
def GetRunTests(self, mode, arch):
|
|
return [ self.GetExecutable(mode, arch, 'run_vm_tests') ]
|
|
|
|
def RunTestCases(cases_to_run, progress, tasks, context):
|
|
progress = PROGRESS_INDICATORS[progress](cases_to_run, context)
|
|
return progress.Run(tasks)
|
|
|
|
|
|
# -------------------------------------------
|
|
# --- T e s t C o n f i g u r a t i o n ---
|
|
# -------------------------------------------
|
|
|
|
class Expression(object):
|
|
pass
|
|
|
|
|
|
class Constant(Expression):
|
|
|
|
def __init__(self, value):
|
|
self.value = value
|
|
|
|
def Evaluate(self, env, defs):
|
|
return self.value
|
|
|
|
|
|
class Variable(Expression):
|
|
|
|
def __init__(self, name):
|
|
self.name = name
|
|
|
|
def GetOutcomes(self, env, defs):
|
|
if self.name in env: return ListSet([env[self.name]])
|
|
else: return Nothing()
|
|
|
|
|
|
class Outcome(Expression):
|
|
|
|
def __init__(self, name):
|
|
self.name = name
|
|
|
|
def GetOutcomes(self, env, defs):
|
|
if self.name in defs:
|
|
return defs[self.name].GetOutcomes(env, defs)
|
|
else:
|
|
return ListSet([self.name])
|
|
|
|
|
|
class Set(object):
|
|
pass
|
|
|
|
|
|
class ListSet(Set):
|
|
|
|
def __init__(self, elms):
|
|
self.elms = elms
|
|
|
|
def __str__(self):
|
|
return "ListSet%s" % str(self.elms)
|
|
|
|
def Intersect(self, that):
|
|
if not isinstance(that, ListSet):
|
|
return that.Intersect(self)
|
|
return ListSet([ x for x in self.elms if x in that.elms ])
|
|
|
|
def Union(self, that):
|
|
if not isinstance(that, ListSet):
|
|
return that.Union(self)
|
|
return ListSet(self.elms + [ x for x in that.elms if x not in self.elms ])
|
|
|
|
def IsEmpty(self):
|
|
return len(self.elms) == 0
|
|
|
|
|
|
class Everything(Set):
|
|
|
|
def Intersect(self, that):
|
|
return that
|
|
|
|
def Union(self, that):
|
|
return self
|
|
|
|
def IsEmpty(self):
|
|
return False
|
|
|
|
|
|
class Nothing(Set):
|
|
|
|
def Intersect(self, that):
|
|
return self
|
|
|
|
def Union(self, that):
|
|
return that
|
|
|
|
def IsEmpty(self):
|
|
return True
|
|
|
|
|
|
class Operation(Expression):
|
|
|
|
def __init__(self, left, op, right):
|
|
self.left = left
|
|
self.op = op
|
|
self.right = right
|
|
|
|
def Evaluate(self, env, defs):
|
|
if self.op == '||' or self.op == ',':
|
|
return self.left.Evaluate(env, defs) or self.right.Evaluate(env, defs)
|
|
elif self.op == 'if':
|
|
return False
|
|
elif self.op == '==':
|
|
inter = self.left.GetOutcomes(env, defs).Intersect(self.right.GetOutcomes(env, defs))
|
|
return not inter.IsEmpty()
|
|
else:
|
|
assert self.op == '&&'
|
|
return self.left.Evaluate(env, defs) and self.right.Evaluate(env, defs)
|
|
|
|
def GetOutcomes(self, env, defs):
|
|
if self.op == '||' or self.op == ',':
|
|
return self.left.GetOutcomes(env, defs).Union(self.right.GetOutcomes(env, defs))
|
|
elif self.op == 'if':
|
|
if self.right.Evaluate(env, defs): return self.left.GetOutcomes(env, defs)
|
|
else: return Nothing()
|
|
else:
|
|
assert self.op == '&&'
|
|
return self.left.GetOutcomes(env, defs).Intersect(self.right.GetOutcomes(env, defs))
|
|
|
|
|
|
def IsAlpha(str):
|
|
for char in str:
|
|
if not (char.isalpha() or char.isdigit() or char == '_'):
|
|
return False
|
|
return True
|
|
|
|
|
|
class Tokenizer(object):
|
|
"""A simple string tokenizer that chops expressions into variables,
|
|
parens and operators"""
|
|
|
|
def __init__(self, expr):
|
|
self.index = 0
|
|
self.expr = expr
|
|
self.length = len(expr)
|
|
self.tokens = None
|
|
|
|
def Current(self, length = 1):
|
|
if not self.HasMore(length): return ""
|
|
return self.expr[self.index:self.index+length]
|
|
|
|
def HasMore(self, length = 1):
|
|
return self.index < self.length + (length - 1)
|
|
|
|
def Advance(self, count = 1):
|
|
self.index = self.index + count
|
|
|
|
def AddToken(self, token):
|
|
self.tokens.append(token)
|
|
|
|
def SkipSpaces(self):
|
|
while self.HasMore() and self.Current().isspace():
|
|
self.Advance()
|
|
|
|
def Tokenize(self):
|
|
self.tokens = [ ]
|
|
while self.HasMore():
|
|
self.SkipSpaces()
|
|
if not self.HasMore():
|
|
return None
|
|
if self.Current() == '(':
|
|
self.AddToken('(')
|
|
self.Advance()
|
|
elif self.Current() == ')':
|
|
self.AddToken(')')
|
|
self.Advance()
|
|
elif self.Current() == '$':
|
|
self.AddToken('$')
|
|
self.Advance()
|
|
elif self.Current() == ',':
|
|
self.AddToken(',')
|
|
self.Advance()
|
|
elif IsAlpha(self.Current()):
|
|
buf = ""
|
|
while self.HasMore() and IsAlpha(self.Current()):
|
|
buf += self.Current()
|
|
self.Advance()
|
|
self.AddToken(buf)
|
|
elif self.Current(2) == '&&':
|
|
self.AddToken('&&')
|
|
self.Advance(2)
|
|
elif self.Current(2) == '||':
|
|
self.AddToken('||')
|
|
self.Advance(2)
|
|
elif self.Current(2) == '==':
|
|
self.AddToken('==')
|
|
self.Advance(2)
|
|
else:
|
|
return None
|
|
return self.tokens
|
|
|
|
|
|
class Scanner(object):
|
|
"""A simple scanner that can serve out tokens from a given list"""
|
|
|
|
def __init__(self, tokens):
|
|
self.tokens = tokens
|
|
self.length = len(tokens)
|
|
self.index = 0
|
|
|
|
def HasMore(self):
|
|
return self.index < self.length
|
|
|
|
def Current(self):
|
|
return self.tokens[self.index]
|
|
|
|
def Advance(self):
|
|
self.index = self.index + 1
|
|
|
|
|
|
def ParseAtomicExpression(scan):
|
|
if scan.Current() == "true":
|
|
scan.Advance()
|
|
return Constant(True)
|
|
elif scan.Current() == "false":
|
|
scan.Advance()
|
|
return Constant(False)
|
|
elif IsAlpha(scan.Current()):
|
|
name = scan.Current()
|
|
scan.Advance()
|
|
return Outcome(name.lower())
|
|
elif scan.Current() == '$':
|
|
scan.Advance()
|
|
if not IsAlpha(scan.Current()):
|
|
return None
|
|
name = scan.Current()
|
|
scan.Advance()
|
|
return Variable(name.lower())
|
|
elif scan.Current() == '(':
|
|
scan.Advance()
|
|
result = ParseLogicalExpression(scan)
|
|
if (not result) or (scan.Current() != ')'):
|
|
return None
|
|
scan.Advance()
|
|
return result
|
|
else:
|
|
return None
|
|
|
|
|
|
BINARIES = ['==']
|
|
def ParseOperatorExpression(scan):
|
|
left = ParseAtomicExpression(scan)
|
|
if not left: return None
|
|
while scan.HasMore() and (scan.Current() in BINARIES):
|
|
op = scan.Current()
|
|
scan.Advance()
|
|
right = ParseOperatorExpression(scan)
|
|
if not right:
|
|
return None
|
|
left = Operation(left, op, right)
|
|
return left
|
|
|
|
|
|
def ParseConditionalExpression(scan):
|
|
left = ParseOperatorExpression(scan)
|
|
if not left: return None
|
|
while scan.HasMore() and (scan.Current() == 'if'):
|
|
scan.Advance()
|
|
right = ParseOperatorExpression(scan)
|
|
if not right:
|
|
return None
|
|
left= Operation(left, 'if', right)
|
|
return left
|
|
|
|
|
|
LOGICALS = ["&&", "||", ","]
|
|
def ParseLogicalExpression(scan):
|
|
left = ParseConditionalExpression(scan)
|
|
if not left: return None
|
|
while scan.HasMore() and (scan.Current() in LOGICALS):
|
|
op = scan.Current()
|
|
scan.Advance()
|
|
right = ParseConditionalExpression(scan)
|
|
if not right:
|
|
return None
|
|
left = Operation(left, op, right)
|
|
return left
|
|
|
|
|
|
def ParseCondition(expr):
|
|
"""Parses a logical expression into an Expression object"""
|
|
tokens = Tokenizer(expr).Tokenize()
|
|
if not tokens:
|
|
print "Malformed expression: '%s'" % expr
|
|
return None
|
|
scan = Scanner(tokens)
|
|
ast = ParseLogicalExpression(scan)
|
|
if not ast:
|
|
print "Malformed expression: '%s'" % expr
|
|
return None
|
|
if scan.HasMore():
|
|
print "Malformed expression: '%s'" % expr
|
|
return None
|
|
return ast
|
|
|
|
|
|
class ClassifiedTest(object):
|
|
|
|
def __init__(self, case, outcomes):
|
|
self.case = case
|
|
self.outcomes = outcomes
|
|
|
|
|
|
class Configuration(object):
|
|
"""The parsed contents of a configuration file"""
|
|
|
|
def __init__(self, sections, defs):
|
|
self.sections = sections
|
|
self.defs = defs
|
|
|
|
def ClassifyTests(self, cases, env):
|
|
sections = [s for s in self.sections if s.condition.Evaluate(env, self.defs)]
|
|
all_rules = reduce(list.__add__, [s.rules for s in sections], [])
|
|
unused_rules = set(all_rules)
|
|
result = [ ]
|
|
all_outcomes = set([])
|
|
for case in cases:
|
|
matches = [ r for r in all_rules if r.Contains(case.path) ]
|
|
outcomes = set([])
|
|
for rule in matches:
|
|
outcomes = outcomes.union(rule.GetOutcomes(env, self.defs))
|
|
unused_rules.discard(rule)
|
|
if not outcomes:
|
|
outcomes = [testing.PASS]
|
|
case.outcomes = outcomes
|
|
all_outcomes = all_outcomes.union(outcomes)
|
|
result.append(ClassifiedTest(case, outcomes))
|
|
return (result, list(unused_rules), all_outcomes)
|
|
|
|
|
|
class Section(object):
|
|
"""A section of the configuration file. Sections are enabled or
|
|
disabled prior to running the tests, based on their conditions"""
|
|
|
|
def __init__(self, condition):
|
|
self.condition = condition
|
|
self.rules = [ ]
|
|
|
|
def AddRule(self, rule):
|
|
self.rules.append(rule)
|
|
|
|
|
|
class Rule(object):
|
|
"""A single rule that specifies the expected outcome for a single
|
|
test."""
|
|
|
|
def __init__(self, raw_path, path, value):
|
|
self.raw_path = raw_path
|
|
self.path = path
|
|
self.value = value
|
|
|
|
def GetOutcomes(self, env, defs):
|
|
set = self.value.GetOutcomes(env, defs)
|
|
assert isinstance(set, ListSet)
|
|
return set.elms
|
|
|
|
def Contains(self, path):
|
|
if len(self.path) > len(path):
|
|
return False
|
|
for i in xrange(len(self.path)):
|
|
if not self.path[i].match(path[i]):
|
|
return False
|
|
return True
|
|
|
|
|
|
HEADER_PATTERN = re.compile(r'\[([^]]+)\]')
|
|
RULE_PATTERN = re.compile(r'\s*([^: ]*)\s*:(.*)')
|
|
DEF_PATTERN = re.compile(r'^def\s*(\w+)\s*=(.*)$')
|
|
PREFIX_PATTERN = re.compile(r'^\s*prefix\s+([\w\_\.\-\/]+)$')
|
|
|
|
|
|
def ReadConfigurationInto(path, sections, defs):
|
|
current_section = Section(Constant(True))
|
|
sections.append(current_section)
|
|
prefix = []
|
|
for line in utils.ReadLinesFrom(path):
|
|
header_match = HEADER_PATTERN.match(line)
|
|
if header_match:
|
|
condition_str = header_match.group(1).strip()
|
|
condition = ParseCondition(condition_str)
|
|
new_section = Section(condition)
|
|
sections.append(new_section)
|
|
current_section = new_section
|
|
continue
|
|
rule_match = RULE_PATTERN.match(line)
|
|
if rule_match:
|
|
path = prefix + SplitPath(rule_match.group(1).strip())
|
|
value_str = rule_match.group(2).strip()
|
|
value = ParseCondition(value_str)
|
|
if not value:
|
|
return False
|
|
current_section.AddRule(Rule(rule_match.group(1), path, value))
|
|
continue
|
|
def_match = DEF_PATTERN.match(line)
|
|
if def_match:
|
|
name = def_match.group(1).lower()
|
|
value = ParseCondition(def_match.group(2).strip())
|
|
if not value:
|
|
return False
|
|
defs[name] = value
|
|
continue
|
|
prefix_match = PREFIX_PATTERN.match(line)
|
|
if prefix_match:
|
|
prefix = SplitPath(prefix_match.group(1).strip())
|
|
continue
|
|
print "Malformed line: '%s'." % line
|
|
return False
|
|
return True
|
|
|
|
|
|
# ---------------
|
|
# --- M a i n ---
|
|
# ---------------
|
|
|
|
|
|
def BuildOptions():
|
|
result = optparse.OptionParser()
|
|
result.add_option("-m", "--mode",
|
|
help="The test modes in which to run (comma-separated)",
|
|
metavar='[all,debug,release]',
|
|
default='debug')
|
|
result.add_option("-v", "--verbose",
|
|
help="Verbose output",
|
|
default=False,
|
|
action="store_true")
|
|
result.add_option("-p", "--progress",
|
|
help="The style of progress indicator (verbose, line, color, mono)",
|
|
choices=PROGRESS_INDICATORS.keys(),
|
|
default=None)
|
|
result.add_option("--report",
|
|
help="Print a summary of the tests to be run",
|
|
default=False,
|
|
action="store_true")
|
|
result.add_option("--list",
|
|
help="List all the tests, but don't run them",
|
|
default=False,
|
|
action="store_true")
|
|
result.add_option("-s", "--suite",
|
|
help="A test suite",
|
|
default=[],
|
|
action="append")
|
|
result.add_option("-t", "--timeout",
|
|
help="Timeout in seconds",
|
|
default=None,
|
|
type="int")
|
|
result.add_option("--checked",
|
|
help="Run tests in checked mode",
|
|
default=False,
|
|
action="store_true")
|
|
result.add_option("--flag",
|
|
help="Pass this additional flag to the VM",
|
|
default=[],
|
|
action="append")
|
|
result.add_option("--arch",
|
|
help="The architecture to run tests for",
|
|
metavar="[all,ia32,x64,simarm,arm,dartc]",
|
|
default=ARCH_GUESS)
|
|
result.add_option("--os",
|
|
help="The OS to run tests on",
|
|
default=OS_GUESS)
|
|
result.add_option("--valgrind",
|
|
help="Run tests through valgrind",
|
|
default=False,
|
|
action="store_true")
|
|
result.add_option("-j", "--tasks",
|
|
help="The number of parallel tasks to run",
|
|
metavar=testing.HOST_CPUS,
|
|
default=testing.USE_DEFAULT_CPUS,
|
|
type="int")
|
|
result.add_option("--time",
|
|
help="Print timing information after running",
|
|
default=False,
|
|
action="store_true")
|
|
result.add_option("--executable",
|
|
help="The executable with which to run the tests",
|
|
default=None)
|
|
result.add_option("--keep_temporary_files",
|
|
help="Do not delete temporary files after running the tests",
|
|
default=False,
|
|
action="store_true")
|
|
result.add_option("--batch",
|
|
help="Run multiple tests for dartc architecture in a single vm",
|
|
choices=["true","false"],
|
|
default="true",
|
|
type="choice");
|
|
result.add_option("--optimize",
|
|
help="Invoke dart compiler with --optimize flag",
|
|
default=False,
|
|
action="store_true")
|
|
|
|
return result
|
|
|
|
|
|
def ProcessOptions(options):
|
|
global VERBOSE
|
|
VERBOSE = options.verbose
|
|
if options.arch == 'all':
|
|
options.arch = 'ia32,x64,simarm'
|
|
if options.mode == 'all':
|
|
options.mode = 'debug,release'
|
|
# By default we run with a higher timeout setting in when running on
|
|
# a simulated architecture and in debug mode.
|
|
if not options.timeout:
|
|
options.timeout = TIMEOUT_SECS
|
|
if 'dartc' in options.arch: options.timeout *= 4
|
|
elif 'chromium' in options.arch: options.timeout *= 4
|
|
elif 'dartium' in options.arch: options.timeout *= 4
|
|
elif 'debug' in options.mode: options.timeout *= 2
|
|
# TODO(zundel): is arch 'sim' out of date?
|
|
if 'sim' in options.arch: options.timeout *= 4
|
|
options.mode = options.mode.split(',')
|
|
options.arch = options.arch.split(',')
|
|
for mode in options.mode:
|
|
if not mode in ['debug', 'release']:
|
|
print "Unknown mode %s" % mode
|
|
return False
|
|
for arch in options.arch:
|
|
if not arch in ['ia32', 'x64', 'simarm', 'arm', 'dartc', 'dartium',
|
|
'chromium']:
|
|
print "Unknown arch %s" % arch
|
|
return False
|
|
options.flags = []
|
|
if (arch == 'dartc' or arch == 'chromium') and mode == 'release':
|
|
options.flags.append('--optimize')
|
|
options.flags.append('--ignore-unrecognized-flags')
|
|
if options.checked:
|
|
options.flags.append('--enable_asserts')
|
|
options.flags.append('--enable_type_checks')
|
|
if options.optimize:
|
|
options.flags.append('--optimize')
|
|
for flag in options.flag:
|
|
options.flags.append(flag)
|
|
if options.verbose:
|
|
print "Flags on the command line:"
|
|
for x in options.flags:
|
|
print x
|
|
# If the user hasn't specified the progress indicator, we pick
|
|
# a good one depending on the setting of the verbose option.
|
|
if not options.progress:
|
|
if options.verbose: options.progress = 'verbose'
|
|
else: options.progress = 'mono'
|
|
# Options for future use. Such as Windows runner support.
|
|
options.suppress_dialogs = True
|
|
options.special_command = None
|
|
return True
|
|
|
|
|
|
REPORT_TEMPLATE = """\
|
|
Total: %(total)i tests
|
|
* %(skipped)4d tests will be skipped
|
|
* %(nocrash)4d tests are expected to be flaky but not crash
|
|
* %(pass)4d tests are expected to pass
|
|
* %(fail_ok)4d tests are expected to fail that we won't fix
|
|
* %(fail)4d tests are expected to fail that we should fix
|
|
* %(crash)4d tests are expected to crash that we should fix
|
|
* %(batched)4d tests are running in batch mode\
|
|
"""
|
|
|
|
def PrintReport(cases):
|
|
"""Print a breakdown of which tests are marked pass/skip/fail """
|
|
def IsFlaky(o):
|
|
return ((testing.PASS in o) and (testing.FAIL in o)
|
|
and (not testing.CRASH in o) and (not testing.OKAY in o))
|
|
def IsFailOk(o):
|
|
return (len(o) == 2) and (testing.FAIL in o) and (testing.OKAY in o)
|
|
unskipped = [c for c in cases if not testing.SKIP in c.outcomes]
|
|
print REPORT_TEMPLATE % {
|
|
'total': len(cases),
|
|
'skipped': len(cases) - len(unskipped),
|
|
'nocrash': len([t for t in unskipped if IsFlaky(t.outcomes)]),
|
|
'pass': len([t for t in unskipped if list(t.outcomes) == [testing.PASS]]),
|
|
'fail_ok': len([t for t in unskipped if IsFailOk(t.outcomes)]),
|
|
'fail': len([t for t in unskipped if list(t.outcomes) == [testing.FAIL]]),
|
|
'crash': len([t for t in unskipped if list(t.outcomes) == [testing.CRASH]]),
|
|
'batched' : len([t for t in unskipped if t.case.IsBatchable()])
|
|
}
|
|
|
|
|
|
def PrintTests(cases):
|
|
has_errors = False
|
|
for case in cases:
|
|
try:
|
|
case.case.GetCommand()
|
|
except:
|
|
sys.stderr.write(case.case.filename + '\n')
|
|
has_errors = True
|
|
if has_errors:
|
|
raise Exception('Errors in above files')
|
|
for case in [c for c in cases if not testing.SKIP in c.outcomes]:
|
|
print "%s\t%s\t%s\t%s" %('/'.join(case.case.path),
|
|
','.join(case.outcomes),
|
|
case.case.IsNegative(),
|
|
'\t'.join(case.case.GetCommand()[1:]))
|
|
|
|
|
|
class Pattern(object):
|
|
|
|
def __init__(self, pattern):
|
|
self.pattern = pattern
|
|
self.compiled = None
|
|
|
|
def match(self, str):
|
|
if not self.compiled:
|
|
pattern = "^" + self.pattern.replace('*', '.*') + "$"
|
|
self.compiled = re.compile(pattern)
|
|
return self.compiled.match(str)
|
|
|
|
def __str__(self):
|
|
return self.pattern
|
|
|
|
|
|
def SplitPath(s):
|
|
stripped = [ c.strip() for c in s.split('/') ]
|
|
return [ Pattern(s) for s in stripped if len(s) > 0 ]
|
|
|
|
|
|
def GetSpecialCommandProcessor(value):
|
|
if (not value) or (value.find('@') == -1):
|
|
def ExpandCommand(args):
|
|
return args
|
|
return ExpandCommand
|
|
else:
|
|
pos = value.find('@')
|
|
import urllib
|
|
prefix = urllib.unquote(value[:pos]).split()
|
|
suffix = urllib.unquote(value[pos+1:]).split()
|
|
def ExpandCommand(args):
|
|
return prefix + args + suffix
|
|
return ExpandCommand
|
|
|
|
|
|
def GetSuites(test_root):
|
|
def IsSuite(path):
|
|
return isdir(path) and exists(join(path, 'testcfg.py'))
|
|
return [ f for f in os.listdir(test_root) if IsSuite(join(test_root, f)) ]
|
|
|
|
|
|
def FormatTime(d):
|
|
millis = round(d * 1000) % 1000
|
|
return time.strftime("%M:%S.", time.gmtime(d)) + ("%03i" % millis)
|
|
|
|
|
|
def Main():
|
|
utils.ConfigureJava()
|
|
parser = BuildOptions()
|
|
(options, args) = parser.parse_args()
|
|
if not ProcessOptions(options):
|
|
parser.print_help()
|
|
return 1
|
|
|
|
client = abspath(join(dirname(sys.argv[0]), '..'))
|
|
repositories = []
|
|
for component in os.listdir(client) + ['.']:
|
|
test_path = join(client, component, 'tests')
|
|
if exists(test_path) and isdir(test_path):
|
|
suites = GetSuites(test_path)
|
|
repositories += [TestRepository(join(test_path, name)) for name in suites]
|
|
repositories += [TestRepository(a) for a in options.suite]
|
|
|
|
root = LiteralTestSuite(repositories)
|
|
if len(args) == 0:
|
|
paths = [SplitPath(t) for t in BUILT_IN_TESTS]
|
|
else:
|
|
paths = [ ]
|
|
for arg in args:
|
|
path = SplitPath(arg)
|
|
paths.append(path)
|
|
|
|
# Check for --valgrind option. If enabled, we overwrite the special
|
|
# command flag with a command that uses the tools/valgrind.py script.
|
|
if options.valgrind:
|
|
run_valgrind = join(client, 'runtime', 'tools', 'valgrind.py')
|
|
options.special_command = "python -u " + run_valgrind + " @"
|
|
|
|
context = Context(client,
|
|
VERBOSE,
|
|
options.os,
|
|
options.timeout,
|
|
GetSpecialCommandProcessor(options.special_command),
|
|
options.suppress_dialogs,
|
|
options.executable,
|
|
options.flags,
|
|
options.keep_temporary_files,
|
|
options.batch)
|
|
|
|
# Get status for tests
|
|
sections = [ ]
|
|
defs = { }
|
|
root.GetTestStatus(context, sections, defs)
|
|
config = Configuration(sections, defs)
|
|
|
|
# List the tests
|
|
all_cases = [ ]
|
|
all_unused = [ ]
|
|
unclassified_tests = [ ]
|
|
globally_unused_rules = None
|
|
for path in paths:
|
|
for mode in options.mode:
|
|
for arch in options.arch:
|
|
env = {
|
|
'mode': mode,
|
|
'system': utils.GuessOS(),
|
|
'arch': arch,
|
|
}
|
|
test_list = root.ListTests([], path, context, mode, arch)
|
|
unclassified_tests += test_list
|
|
(cases, unused_rules, all_outcomes) = config.ClassifyTests(test_list, env)
|
|
if globally_unused_rules is None:
|
|
globally_unused_rules = set(unused_rules)
|
|
else:
|
|
globally_unused_rules = globally_unused_rules.intersection(unused_rules)
|
|
all_cases += cases
|
|
all_unused.append(unused_rules)
|
|
|
|
if options.report:
|
|
PrintReport(all_cases)
|
|
|
|
if options.list:
|
|
PrintTests(all_cases)
|
|
return 0;
|
|
|
|
result = None
|
|
def DoSkip(case):
|
|
return testing.SKIP in case.outcomes or testing.SLOW in case.outcomes
|
|
cases_to_run = [ c for c in all_cases if not DoSkip(c) ]
|
|
if len(cases_to_run) == 0:
|
|
print "No tests to run."
|
|
return 0
|
|
else:
|
|
try:
|
|
start = time.time()
|
|
if RunTestCases(cases_to_run, options.progress, options.tasks,
|
|
context):
|
|
result = 0
|
|
else:
|
|
result = 1
|
|
duration = time.time() - start
|
|
except KeyboardInterrupt:
|
|
print "Exiting on KeyboardInterrupt"
|
|
return 1
|
|
|
|
if options.time:
|
|
print
|
|
print "--- Total time: %s ---" % FormatTime(duration)
|
|
timed_tests = [ t.case for t in cases_to_run if not t.case.duration is None ]
|
|
timed_tests.sort(lambda a, b: a.CompareTime(b))
|
|
index = 1
|
|
for entry in timed_tests[:20]:
|
|
t = FormatTime(entry.duration)
|
|
print "%4i (%s) %s" % (index, t, entry.GetLabel())
|
|
index += 1
|
|
|
|
return result
|
|
|
|
|
|
if __name__ == '__main__':
|
|
sys.exit(Main())
|