mirror of https://github.com/tc39/test262.git
Remove Python-based test runner (#748)
TC-39 agreed to remove this tool from the Test262 repository in May of 2016 [1]. It has since been re-published as a standalone project to facilitate independent development by interested parties [2]. Remove it from the Test262 repository. [1] https://github.com/tc39/tc39-notes/blob/master/es7/2016-05/may-25.md [2] https://github.com/test262-utils/test262-harness-py
This commit is contained in:
parent
cac55f2999
commit
ba32580be5
|
@ -1,126 +0,0 @@
|
|||
## Using the Console Test Runner
|
||||
|
||||
**Note** This utility has been deprecated. It will not be updated with new
|
||||
features; future changes will be limited to bug fixes only.
|
||||
|
||||
The console test runner is used to test browserless implementations of ECMAScript, e.g., [v8](http://en.wikipedia.org/wiki/V8_(JavaScript_engine)), [node](http://en.wikipedia.org/wiki/Node.js), or [js24](http://packages.ubuntu.com/trusty/libmozjs-24-bin)
|
||||
|
||||
### Requirements
|
||||
|
||||
To use the `test262.py` runner, you must have the following:
|
||||
|
||||
* a checkout of the [test262 project](https://github.com/tc39/test262/)
|
||||
* Python 2.7
|
||||
* the Python YAML library [PyYAML](http://www.pyyaml.org)
|
||||
* the javascript engine you intend to test (node, v8, etc.)
|
||||
|
||||
### Quick Start
|
||||
|
||||
To confirm the console test runner is working on a UNIX-like system
|
||||
|
||||
```
|
||||
test262$ ./tools/packaging/test262.py --command "node" 7.2_A1.1_T1
|
||||
ch07/7.2/S7.2_A1.1_T1 passed in non-strict mode
|
||||
|
||||
test262$
|
||||
```
|
||||
|
||||
On a Windows system:
|
||||
|
||||
```
|
||||
Z:\test262>tools\packaging\test262.py --command="node" 7.2_A1.1_T1
|
||||
ch07\7.2\S7.2_A1.1_T1 passed in non-strict mode
|
||||
|
||||
|
||||
Z:\test262>
|
||||
```
|
||||
|
||||
If this does not work, see Troubleshooting (below)
|
||||
|
||||
### Options
|
||||
|
||||
Name | Action
|
||||
-----|-------
|
||||
-h, --help | displays a brief help message
|
||||
--command=COMMAND | **required** command which invokes javascript engine to be tested
|
||||
--tests=TESTS | path to the test suite; default is current directory
|
||||
--cat | don't execute tests, just print code that would be run
|
||||
--summary | generate a summary at end of execution
|
||||
--full-summary | generate a longer summary with details of test failures
|
||||
--strict_only | run only tests that are marked **onlyStrict**
|
||||
--non_strict_only | run only tests that are marked **noStrict**
|
||||
--unmarked_default=MODE | mode to use for tests that are not marked **onlyStrict** or **noStrict** ; MODE can be `strict` or `non_strict` or `both`
|
||||
--logname=LOGNAME | write output to file (in addition to stdout)
|
||||
--junitname=JUNITNAME | write test results to file in JUnit XML format
|
||||
--loglevel=LOGLEVEL | set log level, primarily useful for debugging `test262.py`
|
||||
--print-handle=FUNC | enable async test logging via javascript function e.g., `console.log`
|
||||
|
||||
### Usage Notes
|
||||
|
||||
Non-option arguments are used as filters to match test names. If no filters are found, the whole test suite is run.
|
||||
|
||||
Example | Result
|
||||
-|-
|
||||
test262.py --command="node" | run all tests
|
||||
test262.py --command="node" ch07 ch11 | run tests from chapters 7 and 11
|
||||
test262.py --command="node" 4.4 | run all tests with "4.4" in the name
|
||||
|
||||
The COMMAND argument can be a quoted string. This is useful when testing ECMAScript6 features in node, because node requires the command-line argument `--harmony` to enable ES6:
|
||||
|
||||
```
|
||||
$ test262.py --command="node --harmony" es6
|
||||
```
|
||||
|
||||
#### Async Tests
|
||||
|
||||
Async tests require a 'print' function to be supplied to the test runner. Here are some good defaults:
|
||||
|
||||
Engine | Filename | Print Function
|
||||
-------|----------|---------------
|
||||
V8/Node | node | console.log
|
||||
V8/shell | shell | print
|
||||
SpiderMonkey<sup>1</sup> | js | print
|
||||
JavaScriptCore<sup>2</sup> | jsc | print
|
||||
|
||||
***Notes:***
|
||||
1. As of 2014-Jul-23, SpiderMonkey does not support Promise in the `js` executable ([see bug 911216](https://bugzilla.mozilla.org/show_bug.cgi?id=911216) )
|
||||
2. As of 2014-Jul-23, JavaScriptCore does not support Promise in the `jsc` executable
|
||||
|
||||
|
||||
### Troubleshooting
|
||||
|
||||
#### ImportError: No module named yaml
|
||||
|
||||
On Windows, the message "No module named yaml" can mean that the PyYAML library is installed but not found. If you have this problem, you may be able to use `yaml` interactively from python:
|
||||
|
||||
```
|
||||
Z:\Code\github\test262>python
|
||||
ActivePython 2.7.5.6 (ActiveState Software Inc.) based on
|
||||
Python 2.7.5 (default, Sep 16 2013, 23:16:52) [MSC v.1500 32 bit (Intel)] on win
|
||||
32
|
||||
Type "help", "copyright", "credits" or "license" for more information.
|
||||
>>> import yaml
|
||||
>>> yaml
|
||||
<module 'yaml' from 'C:\Python27\lib\site-packages\yaml\__init__.pyc'>
|
||||
```
|
||||
|
||||
If you can load `yaml` interactively but get the `ImportError` when running `test262.py`:
|
||||
|
||||
```
|
||||
Z:\Code\github\test262>tools\packaging\test262.py --command="node --harmony" --p
|
||||
rint-handle="console.log" ch25
|
||||
Traceback (most recent call last):
|
||||
File "Z:\Code\github\test262\tools\packaging\test262.py", line 31, in <module>
|
||||
|
||||
from parseTestRecord import parseTestRecord, stripHeader
|
||||
File "Z:\Code\github\test262\tools\packaging\parseTestRecord.py", line 20, in
|
||||
<module>
|
||||
import yaml
|
||||
ImportError: No module named yaml
|
||||
```
|
||||
|
||||
Then the fix is to explicitly set the PYTHONPATH environment variable. The location may vary from system to system, but it is typically `'C:\Python27\lib\site-packages`.
|
||||
|
||||
```
|
||||
set PYTHONPATH=C:\Python27\lib\site-packages
|
||||
```
|
|
@ -1,274 +0,0 @@
|
|||
#!/usr/bin/env python
|
||||
|
||||
# Copyright 2014 by Sam Mikes. All rights reserved.
|
||||
# This code is governed by the BSD license found in the LICENSE file.
|
||||
|
||||
import unittest
|
||||
|
||||
import sys
|
||||
import os
|
||||
import cStringIO
|
||||
from functools import wraps
|
||||
|
||||
sys.path.insert(0, "..")
|
||||
|
||||
import test262
|
||||
|
||||
class TestTest262(unittest.TestCase):
|
||||
|
||||
def test_that_tests_run(self):
|
||||
self.assertEqual(1 + 2, 3)
|
||||
|
||||
class MockTest(object):
|
||||
|
||||
def __init__(self, name, negative):
|
||||
self.name = name
|
||||
self.negative = negative if negative else False
|
||||
self.strict_mode = False
|
||||
|
||||
def GetName(self):
|
||||
return self.name
|
||||
|
||||
def IsNegative(self):
|
||||
return self.negative
|
||||
|
||||
def GetMode(self):
|
||||
if self.strict_mode:
|
||||
return "strict mode"
|
||||
|
||||
return "non-strict mode"
|
||||
|
||||
class MockResult(object):
|
||||
|
||||
def __init__(self, case):
|
||||
self.case = case
|
||||
|
||||
|
||||
|
||||
class TestTestSuite(unittest.TestCase):
|
||||
|
||||
def test_that_tests_run(self):
|
||||
self.assertEqual(1 + 2, 3)
|
||||
|
||||
def test_create_test_suite(self):
|
||||
test_suite = test262.TestSuite(".",
|
||||
False,
|
||||
False,
|
||||
False,
|
||||
None)
|
||||
self.assertNotEqual(test_suite, None)
|
||||
|
||||
def test_summary(self):
|
||||
test_suite = test262.TestSuite(".",
|
||||
False,
|
||||
False,
|
||||
False,
|
||||
None)
|
||||
|
||||
progress = test262.ProgressIndicator(100)
|
||||
progress.succeeded = 98
|
||||
progress.failed = 2
|
||||
|
||||
result = mute(True)(test_suite.PrintSummary)(progress, None)
|
||||
self.assertEqual("""
|
||||
=== Summary ===
|
||||
- Ran 100 tests
|
||||
- Passed 98 tests (98.0%)
|
||||
- Failed 2 tests (2.0%)
|
||||
""", result)
|
||||
|
||||
def test_summary_logfile(self):
|
||||
test_suite = test262.TestSuite(".",
|
||||
False,
|
||||
False,
|
||||
False,
|
||||
None)
|
||||
|
||||
progress = test262.ProgressIndicator(100)
|
||||
progress.succeeded = 98
|
||||
progress.failed = 2
|
||||
|
||||
fake_log = cStringIO.StringIO()
|
||||
test_suite.logf = fake_log
|
||||
|
||||
result = mute(True)(test_suite.PrintSummary)(progress, True)
|
||||
|
||||
expected_out = """
|
||||
=== Summary ===
|
||||
- Ran 100 tests
|
||||
- Passed 98 tests (98.0%)
|
||||
- Failed 2 tests (2.0%)
|
||||
"""
|
||||
|
||||
expected_log = """=== Summary ===
|
||||
- Ran 100 tests
|
||||
- Passed 98 tests (98.0%)
|
||||
- Failed 2 tests (2.0%)
|
||||
"""
|
||||
self.assertEqual(expected_out, result)
|
||||
self.assertEqual(expected_log, fake_log.getvalue())
|
||||
|
||||
|
||||
def test_summary_withfails(self):
|
||||
test_suite = test262.TestSuite(".",
|
||||
False,
|
||||
False,
|
||||
False,
|
||||
None)
|
||||
|
||||
progress = test262.ProgressIndicator(100)
|
||||
progress.succeeded = 98
|
||||
progress.failed = 2
|
||||
progress.failed_tests = [
|
||||
MockResult(MockTest("foo", False)),
|
||||
MockResult(MockTest("bar", True))
|
||||
]
|
||||
|
||||
result = mute(True)(test_suite.PrintSummary)(progress, None)
|
||||
self.assertEqual("""
|
||||
=== Summary ===
|
||||
- Ran 100 tests
|
||||
- Passed 98 tests (98.0%)
|
||||
- Failed 2 tests (2.0%)
|
||||
|
||||
Failed Tests
|
||||
foo in non-strict mode
|
||||
|
||||
Expected to fail but passed ---
|
||||
bar in non-strict mode
|
||||
""", result)
|
||||
|
||||
|
||||
def test_summary_withfails_andlog(self):
|
||||
test_suite = test262.TestSuite(".",
|
||||
False,
|
||||
False,
|
||||
False,
|
||||
None)
|
||||
|
||||
progress = test262.ProgressIndicator(100)
|
||||
progress.succeeded = 98
|
||||
progress.failed = 2
|
||||
progress.failed_tests = [
|
||||
MockResult(MockTest("foo", False)),
|
||||
MockResult(MockTest("bar", True))
|
||||
]
|
||||
|
||||
fake_log = cStringIO.StringIO()
|
||||
test_suite.logf = fake_log
|
||||
|
||||
expected_out = """
|
||||
=== Summary ===
|
||||
- Ran 100 tests
|
||||
- Passed 98 tests (98.0%)
|
||||
- Failed 2 tests (2.0%)
|
||||
|
||||
Failed Tests
|
||||
foo in non-strict mode
|
||||
|
||||
Expected to fail but passed ---
|
||||
bar in non-strict mode
|
||||
"""
|
||||
expected_log = """=== Summary ===
|
||||
- Ran 100 tests
|
||||
- Passed 98 tests (98.0%)
|
||||
- Failed 2 tests (2.0%)
|
||||
Failed Tests
|
||||
foo in non-strict mode
|
||||
Expected to fail but passed ---
|
||||
bar in non-strict mode
|
||||
"""
|
||||
|
||||
result = mute(True)(test_suite.PrintSummary)(progress, True)
|
||||
self.assertEqual(expected_out, result)
|
||||
self.assertEqual(expected_log, fake_log.getvalue())
|
||||
|
||||
|
||||
def test_summary_success_logfile(self):
|
||||
test_suite = test262.TestSuite(".",
|
||||
False,
|
||||
False,
|
||||
False,
|
||||
None)
|
||||
|
||||
progress = test262.ProgressIndicator(100)
|
||||
progress.succeeded = 100
|
||||
progress.failed = 0
|
||||
|
||||
fake_log = cStringIO.StringIO()
|
||||
test_suite.logf = fake_log
|
||||
|
||||
result = mute(True)(test_suite.PrintSummary)(progress, True)
|
||||
|
||||
expected_out = """
|
||||
=== Summary ===
|
||||
- Ran 100 tests
|
||||
- All tests succeeded
|
||||
"""
|
||||
|
||||
expected_log = """=== Summary ===
|
||||
- Ran 100 tests
|
||||
- All tests succeeded
|
||||
"""
|
||||
self.assertEqual(expected_out, result)
|
||||
self.assertEqual(expected_log, fake_log.getvalue())
|
||||
|
||||
|
||||
def test_percent_format(self):
|
||||
self.assertEqual(test262.PercentFormat(1, 100), "1 test (1.0%)")
|
||||
self.assertEqual(test262.PercentFormat(0, 100), "0 tests (0.0%)")
|
||||
self.assertEqual(test262.PercentFormat(99, 100), "99 tests (99.0%)")
|
||||
|
||||
|
||||
# module level utility functions
|
||||
# copied from https://stackoverflow.com/questions/2828953/silence-the-stdout-of-a-function-in-python-without-trashing-sys-stdout-and-resto
|
||||
|
||||
|
||||
def mute(returns_output=False):
|
||||
"""
|
||||
Decorate a function that prints to stdout, intercepting the output.
|
||||
If "returns_output" is True, the function will return a generator
|
||||
yielding the printed lines instead of the return values.
|
||||
|
||||
The decorator litterally hijack sys.stdout during each function
|
||||
execution for ALL THE THREADS, so be careful with what you apply it to
|
||||
and in which context.
|
||||
|
||||
>>> def numbers():
|
||||
print "42"
|
||||
print "1984"
|
||||
...
|
||||
>>> numbers()
|
||||
42
|
||||
1984
|
||||
>>> mute()(numbers)()
|
||||
>>> list(mute(True)(numbers)())
|
||||
['42', '1984']
|
||||
|
||||
"""
|
||||
|
||||
def decorator(func):
|
||||
|
||||
@wraps(func)
|
||||
def wrapper(*args, **kwargs):
|
||||
|
||||
saved_stdout = sys.stdout
|
||||
sys.stdout = cStringIO.StringIO()
|
||||
|
||||
try:
|
||||
out = func(*args, **kwargs)
|
||||
if returns_output:
|
||||
out = sys.stdout.getvalue()
|
||||
finally:
|
||||
sys.stdout = saved_stdout
|
||||
|
||||
return out
|
||||
|
||||
return wrapper
|
||||
|
||||
return decorator
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
unittest.main()
|
||||
|
|
@ -1,646 +0,0 @@
|
|||
#!/usr/bin/env python
|
||||
# Copyright 2009 the Sputnik authors. All rights reserved.
|
||||
# This code is governed by the BSD license found in the LICENSE file.
|
||||
|
||||
# This is derived from sputnik.py, the Sputnik console test runner,
|
||||
# with elements from packager.py, which is separately
|
||||
# copyrighted. TODO: Refactor so there is less duplication between
|
||||
# test262.py and packager.py.
|
||||
|
||||
|
||||
import logging
|
||||
import optparse
|
||||
import os
|
||||
from os import path
|
||||
import platform
|
||||
import re
|
||||
import subprocess
|
||||
import sys
|
||||
import tempfile
|
||||
import time
|
||||
import xml.dom.minidom
|
||||
import datetime
|
||||
import shutil
|
||||
import json
|
||||
import stat
|
||||
import xml.etree.ElementTree as xmlj
|
||||
import unicodedata
|
||||
from collections import Counter
|
||||
|
||||
|
||||
from parseTestRecord import parseTestRecord, stripHeader
|
||||
|
||||
from packagerConfig import *
|
||||
|
||||
class Test262Error(Exception):
|
||||
def __init__(self, message):
|
||||
self.message = message
|
||||
|
||||
def ReportError(s):
|
||||
raise Test262Error(s)
|
||||
|
||||
|
||||
|
||||
if not os.path.exists(EXCLUDED_FILENAME):
|
||||
print "Cannot generate (JSON) test262 tests without a file," + \
|
||||
" %s, showing which tests have been disabled!" % EXCLUDED_FILENAME
|
||||
sys.exit(1)
|
||||
EXCLUDE_LIST = xml.dom.minidom.parse(EXCLUDED_FILENAME)
|
||||
EXCLUDE_REASON = EXCLUDE_LIST.getElementsByTagName("reason")
|
||||
EXCLUDE_LIST = EXCLUDE_LIST.getElementsByTagName("test")
|
||||
EXCLUDE_LIST = [x.getAttribute("id") for x in EXCLUDE_LIST]
|
||||
|
||||
|
||||
def BuildOptions():
|
||||
result = optparse.OptionParser()
|
||||
result.add_option("--command", default=None, help="The command-line to run")
|
||||
result.add_option("--tests", default=path.abspath('.'),
|
||||
help="Path to the tests")
|
||||
result.add_option("--cat", default=False, action="store_true",
|
||||
help="Print packaged test code that would be run")
|
||||
result.add_option("--summary", default=False, action="store_true",
|
||||
help="Print summary after running tests")
|
||||
result.add_option("--full-summary", default=False, action="store_true",
|
||||
help="Print summary and test output after running tests")
|
||||
result.add_option("--strict_only", default=False, action="store_true",
|
||||
help="Test only strict mode")
|
||||
result.add_option("--non_strict_only", default=False, action="store_true",
|
||||
help="Test only non-strict mode")
|
||||
result.add_option("--unmarked_default", default="both",
|
||||
help="default mode for tests of unspecified strictness")
|
||||
result.add_option("--logname", help="Filename to save stdout to")
|
||||
result.add_option("--junitname", help="Filename to save test results in JUnit XML format")
|
||||
result.add_option("--loglevel", default="warning",
|
||||
help="sets log level to debug, info, warning, error, or critical")
|
||||
result.add_option("--print-handle", default="print", help="Command to print from console")
|
||||
result.add_option("--list-includes", default=False, action="store_true",
|
||||
help="List includes required by tests")
|
||||
return result
|
||||
|
||||
|
||||
def ValidateOptions(options):
|
||||
if not options.command:
|
||||
ReportError("A --command must be specified.")
|
||||
if not path.exists(options.tests):
|
||||
ReportError("Couldn't find test path '%s'" % options.tests)
|
||||
|
||||
|
||||
placeHolderPattern = re.compile(r"\{\{(\w+)\}\}")
|
||||
|
||||
|
||||
def IsWindows():
|
||||
p = platform.system()
|
||||
return (p == 'Windows') or (p == 'Microsoft')
|
||||
|
||||
|
||||
class TempFile(object):
|
||||
|
||||
def __init__(self, suffix="", prefix="tmp", text=False):
|
||||
self.suffix = suffix
|
||||
self.prefix = prefix
|
||||
self.text = text
|
||||
self.fd = None
|
||||
self.name = None
|
||||
self.is_closed = False
|
||||
self.Open()
|
||||
|
||||
def Open(self):
|
||||
(self.fd, self.name) = tempfile.mkstemp(
|
||||
suffix = self.suffix,
|
||||
prefix = self.prefix,
|
||||
text = self.text)
|
||||
|
||||
def Write(self, str):
|
||||
os.write(self.fd, str)
|
||||
|
||||
def Read(self):
|
||||
f = file(self.name)
|
||||
result = f.read()
|
||||
f.close()
|
||||
return result
|
||||
|
||||
def Close(self):
|
||||
if not self.is_closed:
|
||||
self.is_closed = True
|
||||
os.close(self.fd)
|
||||
|
||||
def Dispose(self):
|
||||
try:
|
||||
self.Close()
|
||||
os.unlink(self.name)
|
||||
except OSError, e:
|
||||
logging.error("Error disposing temp file: %s", str(e))
|
||||
|
||||
|
||||
class TestResult(object):
|
||||
|
||||
def __init__(self, exit_code, stdout, stderr, case):
|
||||
self.exit_code = exit_code
|
||||
self.stdout = stdout
|
||||
self.stderr = stderr
|
||||
self.case = case
|
||||
|
||||
def ReportOutcome(self, long_format):
|
||||
name = self.case.GetName()
|
||||
mode = self.case.GetMode()
|
||||
if self.HasUnexpectedOutcome():
|
||||
if self.case.IsNegative():
|
||||
print "=== %s was expected to fail in %s, but didn't ===" % (name, mode)
|
||||
print "--- expected error: %s ---\n" % self.case.GetNegative()
|
||||
else:
|
||||
if long_format:
|
||||
print "=== %s failed in %s ===" % (name, mode)
|
||||
else:
|
||||
print "%s in %s: " % (name, mode)
|
||||
self.WriteOutput(sys.stdout)
|
||||
if long_format:
|
||||
print "==="
|
||||
elif self.case.IsNegative():
|
||||
print "%s failed in %s as expected" % (name, mode)
|
||||
else:
|
||||
print "%s passed in %s" % (name, mode)
|
||||
|
||||
def WriteOutput(self, target):
|
||||
out = self.stdout.strip()
|
||||
if len(out) > 0:
|
||||
target.write("--- output --- \n %s" % out)
|
||||
err = self.stderr.strip()
|
||||
if len(err) > 0:
|
||||
target.write("--- errors --- \n %s" % err)
|
||||
|
||||
# This is a way to make the output from the "whitespace" tests into valid XML
|
||||
def SafeFormat(self, msg):
|
||||
try:
|
||||
msg = msg.encode(encoding='ascii', errors='strict')
|
||||
msg = msg.replace('\u000Bx', '?')
|
||||
msg = msg.replace('\u000Cx', '?')
|
||||
except:
|
||||
return 'Output contained invalid characters'
|
||||
|
||||
def XmlAssemble(self, result):
|
||||
test_name = self.case.GetName()
|
||||
test_mode = self.case.GetMode()
|
||||
testCaseElement = xmlj.Element("testcase")
|
||||
testpath = self.TestPathManipulation(test_name)
|
||||
testCaseElement.attrib["classname"] = "%s.%s" % (testpath[0] , testpath[1])
|
||||
testCaseElement.attrib["name"] = "%s %s" % (testpath[2].replace('.','_') , test_mode)
|
||||
if self.HasUnexpectedOutcome():
|
||||
failureElement = xmlj.Element("failure")
|
||||
out = self.stdout.strip().decode('utf-8')
|
||||
err = self.stderr.strip().decode('utf-8')
|
||||
if len(out) > 0:
|
||||
failureElement.text = self.SafeFormat(out)
|
||||
if len(err) > 0:
|
||||
failureElement.text = self.SafeFormat(err)
|
||||
testCaseElement.append(failureElement)
|
||||
return testCaseElement
|
||||
|
||||
def TestPathManipulation(self, test_name):
|
||||
testdirlist = test_name.split('/')
|
||||
testcase = testdirlist.pop()
|
||||
testclass = testdirlist.pop()
|
||||
testclass = testclass.replace('.','_')
|
||||
if len(testdirlist) >= 1:
|
||||
testpackage = testdirlist.pop(0)
|
||||
else:
|
||||
testpackage = testclass
|
||||
return(testpackage,testclass,testcase)
|
||||
|
||||
def HasFailed(self):
|
||||
return self.exit_code != 0
|
||||
|
||||
def AsyncHasFailed(self):
|
||||
return 'Test262:AsyncTestComplete' not in self.stdout
|
||||
|
||||
def HasUnexpectedOutcome(self):
|
||||
if self.case.IsAsyncTest():
|
||||
return self.AsyncHasFailed() or self.HasFailed()
|
||||
elif self.case.IsNegative():
|
||||
return not (self.HasFailed() and self.case.NegativeMatch(self.GetErrorOutput()))
|
||||
else:
|
||||
return self.HasFailed()
|
||||
|
||||
def GetErrorOutput(self):
|
||||
if len(self.stderr) != 0:
|
||||
return self.stderr
|
||||
return self.stdout
|
||||
|
||||
|
||||
class TestCase(object):
|
||||
|
||||
def __init__(self, suite, name, full_path, strict_mode):
|
||||
self.suite = suite
|
||||
self.name = name
|
||||
self.full_path = full_path
|
||||
self.strict_mode = strict_mode
|
||||
f = open(self.full_path)
|
||||
self.contents = f.read()
|
||||
f.close()
|
||||
testRecord = parseTestRecord(self.contents, name)
|
||||
self.test = testRecord["test"]
|
||||
del testRecord["test"]
|
||||
del testRecord["header"]
|
||||
testRecord.pop("commentary", None) # do not throw if missing
|
||||
self.testRecord = testRecord;
|
||||
|
||||
self.validate()
|
||||
|
||||
def NegativeMatch(self, stderr):
|
||||
neg = re.compile(self.GetNegative())
|
||||
return re.search(neg, stderr)
|
||||
|
||||
def GetNegative(self):
|
||||
return self.testRecord['negative']
|
||||
|
||||
def GetName(self):
|
||||
return path.join(*self.name)
|
||||
|
||||
def GetMode(self):
|
||||
if self.strict_mode:
|
||||
return "strict mode"
|
||||
else:
|
||||
return "non-strict mode"
|
||||
|
||||
def GetPath(self):
|
||||
return self.name
|
||||
|
||||
def IsNegative(self):
|
||||
return 'negative' in self.testRecord
|
||||
|
||||
def IsOnlyStrict(self):
|
||||
return 'onlyStrict' in self.testRecord
|
||||
|
||||
def IsNoStrict(self):
|
||||
return 'noStrict' in self.testRecord or self.IsRaw()
|
||||
|
||||
def IsRaw(self):
|
||||
return 'raw' in self.testRecord
|
||||
|
||||
def IsAsyncTest(self):
|
||||
return 'async' in self.testRecord
|
||||
|
||||
def GetIncludeList(self):
|
||||
if self.testRecord.get('includes'):
|
||||
return self.testRecord['includes']
|
||||
return []
|
||||
|
||||
def GetAdditionalIncludes(self):
|
||||
return '\n'.join([self.suite.GetInclude(include) for include in self.GetIncludeList()])
|
||||
|
||||
def GetSource(self):
|
||||
if self.IsRaw():
|
||||
return self.test
|
||||
|
||||
source = self.suite.GetInclude("sta.js") + \
|
||||
self.suite.GetInclude("cth.js") + \
|
||||
self.suite.GetInclude("assert.js")
|
||||
|
||||
if self.IsAsyncTest():
|
||||
source = source + \
|
||||
self.suite.GetInclude("timer.js") + \
|
||||
self.suite.GetInclude("doneprintHandle.js").replace('print', self.suite.print_handle)
|
||||
|
||||
source = source + \
|
||||
self.GetAdditionalIncludes() + \
|
||||
self.test + '\n'
|
||||
|
||||
if self.strict_mode:
|
||||
source = '"use strict";\nvar strict_mode = true;\n' + source
|
||||
else:
|
||||
# add comment line so line numbers match in both strict and non-strict version
|
||||
source = '//"no strict";\nvar strict_mode = false;\n' + source
|
||||
|
||||
return source
|
||||
|
||||
def InstantiateTemplate(self, template, params):
|
||||
def GetParameter(match):
|
||||
key = match.group(1)
|
||||
return params.get(key, match.group(0))
|
||||
return placeHolderPattern.sub(GetParameter, template)
|
||||
|
||||
def Execute(self, command):
|
||||
if IsWindows():
|
||||
args = '%s' % command
|
||||
else:
|
||||
args = command.split(" ")
|
||||
stdout = TempFile(prefix="test262-out-")
|
||||
stderr = TempFile(prefix="test262-err-")
|
||||
try:
|
||||
logging.info("exec: %s", str(args))
|
||||
process = subprocess.Popen(
|
||||
args,
|
||||
shell = IsWindows(),
|
||||
stdout = stdout.fd,
|
||||
stderr = stderr.fd
|
||||
)
|
||||
code = process.wait()
|
||||
out = stdout.Read()
|
||||
err = stderr.Read()
|
||||
finally:
|
||||
stdout.Dispose()
|
||||
stderr.Dispose()
|
||||
return (code, out, err)
|
||||
|
||||
def RunTestIn(self, command_template, tmp):
|
||||
tmp.Write(self.GetSource())
|
||||
tmp.Close()
|
||||
command = self.InstantiateTemplate(command_template, {
|
||||
'path': tmp.name
|
||||
})
|
||||
(code, out, err) = self.Execute(command)
|
||||
return TestResult(code, out, err, self)
|
||||
|
||||
def Run(self, command_template):
|
||||
tmp = TempFile(suffix=".js", prefix="test262-", text=True)
|
||||
try:
|
||||
result = self.RunTestIn(command_template, tmp)
|
||||
finally:
|
||||
tmp.Dispose()
|
||||
return result
|
||||
|
||||
def Print(self):
|
||||
print self.GetSource()
|
||||
|
||||
def validate(self):
|
||||
flags = self.testRecord.get("flags")
|
||||
|
||||
if not flags:
|
||||
return
|
||||
|
||||
if 'raw' in flags:
|
||||
if 'noStrict' in flags:
|
||||
raise TypeError("The `raw` flag implies the `noStrict` flag")
|
||||
elif 'onlyStrict' in flags:
|
||||
raise TypeError(
|
||||
"The `raw` flag is incompatible with the `onlyStrict` flag")
|
||||
elif len(self.GetIncludeList()) > 0:
|
||||
raise TypeError(
|
||||
"The `raw` flag is incompatible with the `includes` tag")
|
||||
|
||||
class ProgressIndicator(object):
|
||||
|
||||
def __init__(self, count):
|
||||
self.count = count
|
||||
self.succeeded = 0
|
||||
self.failed = 0
|
||||
self.failed_tests = []
|
||||
|
||||
def HasRun(self, result):
|
||||
result.ReportOutcome(True)
|
||||
if result.HasUnexpectedOutcome():
|
||||
self.failed += 1
|
||||
self.failed_tests.append(result)
|
||||
else:
|
||||
self.succeeded += 1
|
||||
|
||||
|
||||
def MakePlural(n):
|
||||
if (n == 1):
|
||||
return (n, "")
|
||||
else:
|
||||
return (n, "s")
|
||||
|
||||
def PercentFormat(partial, total):
|
||||
return "%i test%s (%.1f%%)" % (MakePlural(partial) +
|
||||
((100.0 * partial)/total,))
|
||||
|
||||
|
||||
class TestSuite(object):
|
||||
|
||||
def __init__(self, root, strict_only, non_strict_only, unmarked_default, print_handle):
|
||||
# TODO: derive from packagerConfig.py
|
||||
self.test_root = path.join(root, 'test')
|
||||
self.lib_root = path.join(root, 'harness')
|
||||
self.strict_only = strict_only
|
||||
self.non_strict_only = non_strict_only
|
||||
self.unmarked_default = unmarked_default
|
||||
self.print_handle = print_handle
|
||||
self.include_cache = { }
|
||||
|
||||
|
||||
def Validate(self):
|
||||
if not path.exists(self.test_root):
|
||||
ReportError("No test repository found")
|
||||
if not path.exists(self.lib_root):
|
||||
ReportError("No test library found")
|
||||
|
||||
def IsHidden(self, path):
|
||||
return path.startswith('.') or path == 'CVS'
|
||||
|
||||
def IsTestCase(self, path):
|
||||
return path.endswith('.js')
|
||||
|
||||
def ShouldRun(self, rel_path, tests):
|
||||
if len(tests) == 0:
|
||||
return True
|
||||
for test in tests:
|
||||
if test in rel_path:
|
||||
return True
|
||||
return False
|
||||
|
||||
def GetInclude(self, name):
|
||||
if not name in self.include_cache:
|
||||
static = path.join(self.lib_root, name)
|
||||
if path.exists(static):
|
||||
f = open(static)
|
||||
contents = stripHeader(f.read())
|
||||
contents = re.sub(r'\r\n', '\n', contents)
|
||||
self.include_cache[name] = contents + "\n"
|
||||
f.close()
|
||||
else:
|
||||
ReportError("Can't find: " + static)
|
||||
return self.include_cache[name]
|
||||
|
||||
def EnumerateTests(self, tests):
|
||||
logging.info("Listing tests in %s", self.test_root)
|
||||
cases = []
|
||||
for root, dirs, files in os.walk(self.test_root):
|
||||
for f in [x for x in dirs if self.IsHidden(x)]:
|
||||
dirs.remove(f)
|
||||
dirs.sort()
|
||||
for f in sorted(files):
|
||||
if self.IsTestCase(f):
|
||||
full_path = path.join(root, f)
|
||||
if full_path.startswith(self.test_root):
|
||||
rel_path = full_path[len(self.test_root)+1:]
|
||||
else:
|
||||
logging.warning("Unexpected path %s", full_path)
|
||||
rel_path = full_path
|
||||
if self.ShouldRun(rel_path, tests):
|
||||
basename = path.basename(full_path)[:-3]
|
||||
name = rel_path.split(path.sep)[:-1] + [basename]
|
||||
if EXCLUDE_LIST.count(basename) >= 1:
|
||||
print 'Excluded: ' + basename
|
||||
else:
|
||||
if not self.non_strict_only:
|
||||
strict_case = TestCase(self, name, full_path, True)
|
||||
if not strict_case.IsNoStrict():
|
||||
if strict_case.IsOnlyStrict() or \
|
||||
self.unmarked_default in ['both', 'strict']:
|
||||
cases.append(strict_case)
|
||||
if not self.strict_only:
|
||||
non_strict_case = TestCase(self, name, full_path, False)
|
||||
if not non_strict_case.IsOnlyStrict():
|
||||
if non_strict_case.IsNoStrict() or \
|
||||
self.unmarked_default in ['both', 'non_strict']:
|
||||
cases.append(non_strict_case)
|
||||
logging.info("Done listing tests")
|
||||
return cases
|
||||
|
||||
|
||||
def PrintSummary(self, progress, logfile):
|
||||
|
||||
def write(s):
|
||||
if logfile:
|
||||
self.logf.write(s + "\n")
|
||||
print s
|
||||
|
||||
print
|
||||
write("=== Summary ===");
|
||||
count = progress.count
|
||||
succeeded = progress.succeeded
|
||||
failed = progress.failed
|
||||
write(" - Ran %i test%s" % MakePlural(count))
|
||||
if progress.failed == 0:
|
||||
write(" - All tests succeeded")
|
||||
else:
|
||||
write(" - Passed " + PercentFormat(succeeded, count))
|
||||
write(" - Failed " + PercentFormat(failed, count))
|
||||
positive = [c for c in progress.failed_tests if not c.case.IsNegative()]
|
||||
negative = [c for c in progress.failed_tests if c.case.IsNegative()]
|
||||
if len(positive) > 0:
|
||||
print
|
||||
write("Failed Tests")
|
||||
for result in positive:
|
||||
write(" %s in %s" % (result.case.GetName(), result.case.GetMode()))
|
||||
if len(negative) > 0:
|
||||
print
|
||||
write("Expected to fail but passed ---")
|
||||
for result in negative:
|
||||
write(" %s in %s" % (result.case.GetName(), result.case.GetMode()))
|
||||
|
||||
def PrintFailureOutput(self, progress, logfile):
|
||||
for result in progress.failed_tests:
|
||||
if logfile:
|
||||
self.WriteLog(result)
|
||||
print
|
||||
result.ReportOutcome(False)
|
||||
|
||||
def Run(self, command_template, tests, print_summary, full_summary, logname, junitfile):
|
||||
if not "{{path}}" in command_template:
|
||||
command_template += " {{path}}"
|
||||
cases = self.EnumerateTests(tests)
|
||||
if len(cases) == 0:
|
||||
ReportError("No tests to run")
|
||||
progress = ProgressIndicator(len(cases))
|
||||
if logname:
|
||||
self.logf = open(logname, "w")
|
||||
if junitfile:
|
||||
self.outfile = open(junitfile, "w")
|
||||
TestSuitesElement = xmlj.Element("testsuites")
|
||||
TestSuiteElement = xmlj.Element("testsuite")
|
||||
TestSuitesElement.append(TestSuiteElement)
|
||||
TestSuiteElement.attrib["name "] = "test262"
|
||||
for x in range(len(EXCLUDE_LIST)):
|
||||
if self.ShouldRun (unicode(EXCLUDE_LIST[x].encode('utf-8','ignore')), tests):
|
||||
SkipCaseElement = xmlj.Element("testcase")
|
||||
SkipCaseElement.attrib["classname"] = unicode(EXCLUDE_LIST[x]).encode('utf-8','ignore')
|
||||
SkipCaseElement.attrib["name"] = unicode(EXCLUDE_LIST[x]).encode('utf-8','ignore')
|
||||
SkipElement = xmlj.Element("skipped")
|
||||
SkipElement.attrib["message"] = unicode(EXCLUDE_REASON[x].firstChild.nodeValue)
|
||||
SkipCaseElement.append(SkipElement)
|
||||
TestSuiteElement.append(SkipCaseElement)
|
||||
|
||||
for case in cases:
|
||||
result = case.Run(command_template)
|
||||
if junitfile:
|
||||
TestCaseElement = result.XmlAssemble(result)
|
||||
TestSuiteElement.append(TestCaseElement)
|
||||
if case == cases[len(cases)-1]:
|
||||
xmlj.ElementTree(TestSuitesElement).write(junitfile, "UTF-8")
|
||||
if logname:
|
||||
self.WriteLog(result)
|
||||
progress.HasRun(result)
|
||||
|
||||
if print_summary:
|
||||
self.PrintSummary(progress, logname)
|
||||
if full_summary:
|
||||
self.PrintFailureOutput(progress, logname)
|
||||
else:
|
||||
print
|
||||
print "Use --full-summary to see output from failed tests"
|
||||
print
|
||||
return progress.failed
|
||||
|
||||
def WriteLog(self, result):
|
||||
name = result.case.GetName()
|
||||
mode = result.case.GetMode()
|
||||
if result.HasUnexpectedOutcome():
|
||||
if result.case.IsNegative():
|
||||
self.logf.write("=== %s was expected to fail in %s, but didn't === \n" % (name, mode))
|
||||
self.logf.write("--- expected error: %s ---\n" % result.case.GetNegative())
|
||||
result.WriteOutput(self.logf)
|
||||
else:
|
||||
self.logf.write("=== %s failed in %s === \n" % (name, mode))
|
||||
result.WriteOutput(self.logf)
|
||||
self.logf.write("===\n")
|
||||
elif result.case.IsNegative():
|
||||
self.logf.write("%s failed in %s as expected \n" % (name, mode))
|
||||
else:
|
||||
self.logf.write("%s passed in %s \n" % (name, mode))
|
||||
|
||||
def Print(self, tests):
|
||||
cases = self.EnumerateTests(tests)
|
||||
if len(cases) > 0:
|
||||
cases[0].Print()
|
||||
|
||||
def ListIncludes(self, tests):
|
||||
cases = self.EnumerateTests(tests)
|
||||
includes_dict = Counter()
|
||||
for case in cases:
|
||||
includes = case.GetIncludeList()
|
||||
includes_dict.update(includes)
|
||||
|
||||
print includes_dict
|
||||
|
||||
|
||||
def Main():
|
||||
code = 0
|
||||
parser = BuildOptions()
|
||||
(options, args) = parser.parse_args()
|
||||
ValidateOptions(options)
|
||||
test_suite = TestSuite(options.tests,
|
||||
options.strict_only,
|
||||
options.non_strict_only,
|
||||
options.unmarked_default,
|
||||
options.print_handle)
|
||||
test_suite.Validate()
|
||||
if options.loglevel == 'debug':
|
||||
logging.basicConfig(level=logging.DEBUG)
|
||||
elif options.loglevel == 'info':
|
||||
logging.basicConfig(level=logging.INFO)
|
||||
elif options.loglevel == 'warning':
|
||||
logging.basicConfig(level=logging.WARNING)
|
||||
elif options.loglevel == 'error':
|
||||
logging.basicConfig(level=logging.ERROR)
|
||||
elif options.loglevel == 'critical':
|
||||
logging.basicConfig(level=logging.CRITICAL)
|
||||
if options.cat:
|
||||
test_suite.Print(args)
|
||||
elif options.list_includes:
|
||||
test_suite.ListIncludes(args)
|
||||
else:
|
||||
code = test_suite.Run(options.command, args,
|
||||
options.summary or options.full_summary,
|
||||
options.full_summary,
|
||||
options.logname,
|
||||
options.junitname)
|
||||
return code
|
||||
|
||||
if __name__ == '__main__':
|
||||
try:
|
||||
code = Main()
|
||||
sys.exit(code)
|
||||
except Test262Error, e:
|
||||
print "Error: %s" % e.message
|
||||
sys.exit(1)
|
Loading…
Reference in New Issue