/external/chromium_org/mojo/tools/pylib/ |
H A D | mojo_python_tests_runner.py | 137 return set(test.id() for test, _ in result.failures + result.errors)
|
/external/chromium_org/tools/telemetry/telemetry/page/ |
H A D | profile_generator.py | 81 if results.failures:
|
/external/chromium_org/tools/telemetry/telemetry/unittest/ |
H A D | progress_reporter.py | 81 return self.failures + self.errors
|
H A D | json_results.py | 172 for test, error in result.failures + result.errors:
|
/external/lldb/examples/test/ |
H A D | usage-lldb-loggings | 37 Session logs for test failures/errors/unexpected successes will go into directory '2012-08-22-11_36_37' 102 Session logs for test failures/errors/unexpected successes will go into directory '2012-08-22-13_21_46'
|
/external/lldb/test/unittest2/test/ |
H A D | test_skipping.py | 99 self.assertFalse(result.failures)
|
/external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/layout_tests/controllers/ |
H A D | layout_test_runner_unittest.py | 142 test = 'failures/expected/reftest.html' 147 result = TestResult(test_name=test, failures=[test_failures.FailureReftestMismatchDidNotOccur()], reftest_type=['!=']) 153 result = TestResult(test_name=test, failures=[], reftest_type=['=='])
|
H A D | test_result_writer.py | 40 expected_driver_output, failures): 48 for failure in failures: 101 # Filename pieces when writing failures to the test results directory.
|
H A D | layout_test_runner.py | 170 "Exiting early after %d failures." % run_results.unexpected_failures) 328 if result.failures: 330 if any([f.driver_needs_restart() for f in result.failures]): 337 for f in result.failures:
|
/external/antlr/antlr-3.4/gunit/src/main/java/org/antlr/gunit/ |
H A D | gUnitExecutor.java | 72 public List<AbstractTest> failures; field in class:gUnitExecutor 98 failures = new ArrayList<AbstractTest>(); 140 testResultST.setAttribute("failure", failures); 203 failures.add(test); 216 failures.add(test); 222 failures.add(test);
|
/external/valgrind/main/tests/ |
H A D | vg_regtest | 154 my @failures; # List of failed tests 403 push(@failures, sprintf("%-40s ($mid)", "$fullname")); 630 foreach my $failure (@failures) {
|
H A D | vg_regtest.in | 154 my @failures; # List of failed tests 403 push(@failures, sprintf("%-40s ($mid)", "$fullname")); 630 foreach my $failure (@failures) {
|
/external/clang/tools/scan-build/ |
H A D | scan-build | 78 Diag ("Preprocessed versions of these sources were deposited in '$Dir/failures'.\n"); 539 if (scalar(@filesFound) == 0 and ! -e "$Dir/failures") { 553 # Scan the failures directory and use the information in the .info files 555 my @failures; 557 if (-d "$Dir/failures") { 558 opendir(DIR, "$Dir/failures"); 559 @failures = grep { /[.]info.txt$/ && !/attribute_ignored/; } readdir(DIR); 561 opendir(DIR, "$Dir/failures"); 564 foreach my $file (@failures) { 565 open IN, "$Dir/failures/ [all...] |
/external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/layout_tests/views/ |
H A D | printing.py | 112 self._print_expected_results_of_type(run_results, test_expectations.FAIL, "failures", tests_with_result_type_callback) 354 result_message = self._result_message(result.type, result.failures, expected, 374 def _result_message(self, result_type, failures, expected, timing, test_run_time): 380 return ' failed%s (%s)%s' % (exp_string, ', '.join(failure.message() for failure in failures), timing_string)
|
/external/chromium_org/third_party/icu/source/test/intltest/ |
H A D | plurfmts.cpp | 607 const char *failures[] = { local 614 int len = sizeof(failures)/sizeof(failures[0]); 618 UnicodeString fmt(failures[i], -1, US_INV);
|
/external/chromium_org/third_party/skia/gm/tests/ |
H A D | run.sh | 173 >$JSON_DIR/different-pixels-ignore-some-failures.json 216 # Failures in selftest1 should be ignored, but failures in selftest2 should not. 221 gm_test "--verbose --hierarchy --match selftest1 $CONFIGS -r $GM_INPUTS/json/different-pixels-ignore-some-failures.json" "$GM_OUTPUTS/ignoring-some-failures"
|
/external/chromium_org/tools/grit/grit/ |
H A D | test_suite_all.py | 159 sys.exit(len(test_result.errors) + len(test_result.failures))
|
/external/chromium_org/tools/telemetry/telemetry/ |
H A D | benchmark.py | 100 return len(results.failures)
|
/external/chromium_org/tools/telemetry/telemetry/results/ |
H A D | page_test_results.py | 99 def failures(self): member in class:PageTestResults
|
/external/icu/icu4c/source/test/intltest/ |
H A D | plurfmts.cpp | 607 const char *failures[] = { local 614 int len = sizeof(failures)/sizeof(failures[0]); 618 UnicodeString fmt(failures[i], -1, US_INV);
|
/external/skia/gm/tests/ |
H A D | run.sh | 173 >$JSON_DIR/different-pixels-ignore-some-failures.json 216 # Failures in selftest1 should be ignored, but failures in selftest2 should not. 221 gm_test "--verbose --hierarchy --match selftest1 $CONFIGS -r $GM_INPUTS/json/different-pixels-ignore-some-failures.json" "$GM_OUTPUTS/ignoring-some-failures"
|
/external/chromium_org/third_party/cython/src/ |
H A D | runtests.py | 884 failures, errors = len(result.failures), len(result.errors) 887 if failures == len(result.failures) and errors == len(result.errors): 1050 self.strip_error_results(self.failures) 1052 return (self.failures, self.errors, self.testsRun, 1059 failures, errors, tests_run, output = data 1063 result.failures.extend(failures) 1553 parser.add_option("--no-cleanup-failures", des [all...] |
/external/okhttp/okhttp-tests/src/test/java/com/squareup/okhttp/internal/http/ |
H A D | RouteSelectorTest.java | 463 List<String> failures = new ArrayList<String>(); field in class:RouteSelectorTest.FakeProxySelector 477 failures.add(
|
/external/antlr/antlr-3.4/runtime/Python/ |
H A D | xmlrunner.py | 133 stream.write('<testsuite errors="%(e)d" failures="%(f)d" ' % \ 134 { "e": len(self.errors), "f": len(self.failures) }) 251 self._try_test_run(TestTest, """<testsuite errors="0" failures="0" name="unittest.TestSuite" tests="0" time="0.000"> 265 self._try_test_run(TestTest, """<testsuite errors="0" failures="0" name="unittest.TestSuite" tests="1" time="0.000"> 280 self._try_test_run(TestTest, """<testsuite errors="0" failures="1" name="unittest.TestSuite" tests="1" time="0.000"> 297 self._try_test_run(TestTest, """<testsuite errors="1" failures="0" name="unittest.TestSuite" tests="1" time="0.000"> 314 self._try_test_run(TestTest, """<testsuite errors="0" failures="0" name="unittest.TestSuite" tests="1" time="0.000"> 330 self._try_test_run(TestTest, """<testsuite errors="0" failures="0" name="unittest.TestSuite" tests="1" time="0.000">
|
/external/chromium_org/chrome/test/mini_installer/ |
H A D | test_installer.py | 404 return set(test.id() for test, _ in result.failures + result.errors)
|