summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rwxr-xr-xselftest/format-subunit10
-rw-r--r--selftest/subunithelper.py35
2 files changed, 26 insertions, 19 deletions
diff --git a/selftest/format-subunit b/selftest/format-subunit
index 43cdf096f2..0d16032afa 100755
--- a/selftest/format-subunit
+++ b/selftest/format-subunit
@@ -32,6 +32,7 @@ class PlainFormatter(subunithelper.TestsuiteEnabledTestResult):
def __init__(self, summaryfile, verbose, immediate, statistics,
totaltests=None):
+ super(PlainFormatter, self).__init__()
self.verbose = verbose
self.immediate = immediate
self.statistics = statistics
@@ -124,8 +125,8 @@ class PlainFormatter(subunithelper.TestsuiteEnabledTestResult):
def startTest(self, test):
pass
- def addSuccess(self, test, details=None):
- self.end_test(test.id(), "success", False, details)
+ def addSuccess(self, test):
+ self.end_test(test.id(), "success", False)
def addError(self, test, details=None):
self.end_test(test.id(), "error", True, details)
@@ -150,9 +151,12 @@ class PlainFormatter(subunithelper.TestsuiteEnabledTestResult):
'success': '.'}.get(result, "?(%s)" % result))
return
+ if not self.name in self.test_output:
+ self.test_output[self.name] = ""
+
self.test_output[self.name] += "UNEXPECTED(%s): %s\n" % (result, testname)
if reason is not None:
- self.test_output[self.name] += "REASON: %s\n" % (reason.strip(),)
+ self.test_output[self.name] += "REASON: %s\n" % (reason[1].message.encode("utf-8").strip(),)
if self.immediate and not self.verbose:
print self.test_output[self.name]
diff --git a/selftest/subunithelper.py b/selftest/subunithelper.py
index e89810f9bc..559f5c0c24 100644
--- a/selftest/subunithelper.py
+++ b/selftest/subunithelper.py
@@ -50,7 +50,7 @@ def parse_results(msg_ops, statistics, fh):
name = arg.rstrip()
test = subunit.RemotedTestCase(name)
if name in open_tests:
- msg_ops.addError(open_tests.pop(name), "Test already running")
+ msg_ops.addError(open_tests.pop(name), subunit.RemoteError(u"Test already running"))
msg_ops.startTest(test)
open_tests[name] = test
elif command == "time":
@@ -81,40 +81,43 @@ def parse_results(msg_ops, statistics, fh):
else:
reason += l
+ remote_error = subunit.RemoteError(reason.decode("utf-8"))
+
if not terminated:
statistics['TESTS_ERROR']+=1
- msg_ops.addError(subunit.RemotedTestCase(testname), "reason (%s) interrupted" % result)
+ msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"reason (%s) interrupted" % result))
return 1
else:
reason = None
+ remote_error = subunit.RemoteError(u"No reason specified")
if result in ("success", "successful"):
try:
test = open_tests.pop(testname)
except KeyError:
statistics['TESTS_ERROR']+=1
- msg_ops.addError(subunit.RemotedTestCase(testname), "Test was never started")
+ msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"Test was never started"))
else:
statistics['TESTS_EXPECTED_OK']+=1
- msg_ops.addSuccess(test, reason)
+ msg_ops.addSuccess(test)
elif result in ("xfail", "knownfail"):
try:
test = open_tests.pop(testname)
except KeyError:
statistics['TESTS_ERROR']+=1
- msg_ops.addError(subunit.RemotedTestCase(testname), "Test was never started")
+ msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"Test was never started"))
else:
statistics['TESTS_EXPECTED_FAIL']+=1
- msg_ops.addExpectedFail(test, reason)
+ msg_ops.addExpectedFailure(test, remote_error)
expected_fail+=1
elif result in ("failure", "fail"):
try:
test = open_tests.pop(testname)
except KeyError:
statistics['TESTS_ERROR']+=1
- msg_ops.addError(subunit.RemotedTestCase(testname), "Test was never started")
+ msg_ops.addError(subunit.RemotedTestCase(testname), subunit.RemoteError(u"Test was never started"))
else:
statistics['TESTS_UNEXPECTED_FAIL']+=1
- msg_ops.addFailure(test, reason)
+ msg_ops.addFailure(test, remote_error)
elif result == "skip":
statistics['TESTS_SKIP']+=1
# Allow tests to be skipped without prior announcement of test
@@ -129,7 +132,7 @@ def parse_results(msg_ops, statistics, fh):
test = open_tests.pop(testname)
except KeyError:
test = subunit.RemotedTestCase(testname)
- msg_ops.addError(test, reason)
+ msg_ops.addError(test, remote_error)
elif result == "skip-testsuite":
msg_ops.skip_testsuite(testname)
elif result == "testsuite-success":
@@ -160,13 +163,13 @@ def parse_results(msg_ops, statistics, fh):
while open_tests:
test = subunit.RemotedTestCase(open_tests.popitem()[1])
- msg_ops.addError(test, "was started but never finished!")
+ msg_ops.addError(test, subunit.RemoteError(u"was started but never finished!"))
statistics['TESTS_ERROR']+=1
if statistics['TESTS_ERROR'] > 0:
return 1
if statistics['TESTS_UNEXPECTED_FAIL'] > 0:
- return 1
+ return 1
return 0
@@ -258,9 +261,9 @@ class FilterOps(testtools.testresult.TestResult):
self._ops.addSkip(test, details)
self.output = None
- def addExpectedFail(self, test, details=None):
+ def addExpectedFailure(self, test, details=None):
test = self._add_prefix(test)
- self._ops.addExpectedFail(test, details)
+ self._ops.addExpectedFailure(test, details)
self.output = None
def addFailure(self, test, details=None):
@@ -270,10 +273,10 @@ class FilterOps(testtools.testresult.TestResult):
self.xfail_added+=1
self.total_xfail+=1
if details is not None:
- details += xfail_reason
+ details = subunit.RemoteError(details[1].message + xfail_reason.decode("utf-8"))
else:
- details = xfail_reason
- self._ops.addExpectedFail(test, details)
+ details = subunit.RemoteError(xfail_reason.decode("utf-8"))
+ self._ops.addExpectedFailure(test, details)
else:
self.fail_added+=1
self.total_fail+=1