/usr/lib/python2.7/dist-packages/testtools/testresult/real.py is in python-testtools 2.3.0-3ubuntu2.
This file is owned by root:root, with mode 0o644.
The actual contents of the file can be viewed below.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505 1506 1507 1508 1509 1510 1511 1512 1513 1514 1515 1516 1517 1518 1519 1520 1521 1522 1523 1524 1525 1526 1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625 1626 1627 1628 1629 1630 1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671 1672 1673 1674 1675 1676 1677 1678 1679 1680 1681 1682 1683 1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728 1729 1730 1731 1732 1733 1734 1735 1736 1737 1738 1739 1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772 1773 1774 1775 1776 1777 1778 1779 1780 1781 1782 1783 1784 1785 1786 1787 1788 1789 1790 1791 1792 1793 1794 1795 1796 1797 1798 1799 1800 1801 1802 1803 1804 1805 1806 1807 1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819 1820 1821 1822 1823 1824 1825 1826 1827 1828 1829 1830 1831 1832 1833 1834 1835 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850 1851 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883 1884 1885 1886 1887 1888 1889 1890 1891 1892 1893 1894 1895 1896 1897 1898 1899 1900 1901 1902 1903 1904 1905 1906 1907 1908 1909 1910 1911 1912 1913 1914 1915 1916 1917 1918 1919 1920 1921 1922 1923 1924 1925 1926 1927 1928 1929 1930 1931 1932 1933 1934 1935 1936 1937 1938 1939 1940 1941 1942 1943 1944 1945 1946 1947 1948 1949 1950 1951 1952 1953 1954 1955 1956 1957 1958 1959 1960 1961 1962 1963 1964 1965 1966 1967 1968 1969 1970 1971 1972 1973 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 2017 2018 2019 2020 2021 2022 2023 2024 2025 2026 2027 2028 2029 2030 2031 2032 2033 2034 2035 2036 2037 2038 2039 2040 2041 2042 2043 2044 2045 2046 2047 2048 2049 2050 2051 2052 2053 2054 2055 2056 2057 2058 2059 2060 2061 2062 2063 2064 2065 2066 2067 2068 2069 2070 2071 2072 2073 2074 2075 2076 2077 2078 2079 2080 2081 2082 2083 2084 2085 2086 2087 2088 2089 2090 2091 2092 2093 2094 2095 2096 2097 2098 2099 2100 2101 2102 2103 2104 | # Copyright (c) 2008-2015 testtools developers. See LICENSE for details.
"""Test results and related things."""
__all__ = [
'ExtendedToOriginalDecorator',
'ExtendedToStreamDecorator',
'MultiTestResult',
'ResourcedToStreamDecorator',
'StreamFailFast',
'StreamResult',
'StreamSummary',
'StreamTagger',
'StreamToDict',
'StreamToExtendedDecorator',
'StreamToQueue',
'Tagger',
'TestControl',
'TestResult',
'TestResultDecorator',
'ThreadsafeForwardingResult',
'TimestampingStreamResult',
]
import datetime
import math
from operator import methodcaller
import sys
import unittest
import warnings
from extras import safe_hasattr, try_import, try_imports
parse_mime_type = try_import('mimeparse.parse_mime_type')
Queue = try_imports(['Queue.Queue', 'queue.Queue'])
from testtools.compat import str_is_unicode, _u, _b
from testtools.content import (
Content,
text_content,
TracebackContent,
)
from testtools.content_type import ContentType
from testtools.tags import TagContext
# circular import
# from testtools.testcase import PlaceHolder
PlaceHolder = None
# From http://docs.python.org/library/datetime.html
_ZERO = datetime.timedelta(0)
class UTC(datetime.tzinfo):
"""UTC"""
def utcoffset(self, dt):
return _ZERO
def tzname(self, dt):
return "UTC"
def dst(self, dt):
return _ZERO
utc = UTC()
class TestResult(unittest.TestResult):
"""Subclass of unittest.TestResult extending the protocol for flexibility.
This test result supports an experimental protocol for providing additional
data to in test outcomes. All the outcome methods take an optional dict
'details'. If supplied any other detail parameters like 'err' or 'reason'
should not be provided. The details dict is a mapping from names to
MIME content objects (see testtools.content). This permits attaching
tracebacks, log files, or even large objects like databases that were
part of the test fixture. Until this API is accepted into upstream
Python it is considered experimental: it may be replaced at any point
by a newer version more in line with upstream Python. Compatibility would
be aimed for in this case, but may not be possible.
:ivar skip_reasons: A dict of skip-reasons -> list of tests. See addSkip.
"""
def __init__(self, failfast=False, tb_locals=False):
# startTestRun resets all attributes, and older clients don't know to
# call startTestRun, so it is called once here.
# Because subclasses may reasonably not expect this, we call the
# specific version we want to run.
self.failfast = failfast
self.tb_locals = tb_locals
TestResult.startTestRun(self)
def addExpectedFailure(self, test, err=None, details=None):
"""Called when a test has failed in an expected manner.
Like with addSuccess and addError, testStopped should still be called.
:param test: The test that has been skipped.
:param err: The exc_info of the error that was raised.
:return: None
"""
# This is the python 2.7 implementation
self.expectedFailures.append(
(test, self._err_details_to_string(test, err, details)))
def addError(self, test, err=None, details=None):
"""Called when an error has occurred. 'err' is a tuple of values as
returned by sys.exc_info().
:param details: Alternative way to supply details about the outcome.
see the class docstring for more information.
"""
self.errors.append(
(test, self._err_details_to_string(test, err, details)))
if self.failfast:
self.stop()
def addFailure(self, test, err=None, details=None):
"""Called when an error has occurred. 'err' is a tuple of values as
returned by sys.exc_info().
:param details: Alternative way to supply details about the outcome.
see the class docstring for more information.
"""
self.failures.append(
(test, self._err_details_to_string(test, err, details)))
if self.failfast:
self.stop()
def addSkip(self, test, reason=None, details=None):
"""Called when a test has been skipped rather than running.
Like with addSuccess and addError, testStopped should still be called.
This must be called by the TestCase. 'addError' and 'addFailure' will
not call addSkip, since they have no assumptions about the kind of
errors that a test can raise.
:param test: The test that has been skipped.
:param reason: The reason for the test being skipped. For instance,
u"pyGL is not available".
:param details: Alternative way to supply details about the outcome.
see the class docstring for more information.
:return: None
"""
if reason is None:
reason = details.get('reason')
if reason is None:
reason = 'No reason given'
else:
reason = reason.as_text()
skip_list = self.skip_reasons.setdefault(reason, [])
skip_list.append(test)
def addSuccess(self, test, details=None):
"""Called when a test succeeded."""
def addUnexpectedSuccess(self, test, details=None):
"""Called when a test was expected to fail, but succeed."""
self.unexpectedSuccesses.append(test)
if self.failfast:
self.stop()
def wasSuccessful(self):
"""Has this result been successful so far?
If there have been any errors, failures or unexpected successes,
return False. Otherwise, return True.
Note: This differs from standard unittest in that we consider
unexpected successes to be equivalent to failures, rather than
successes.
"""
return not (self.errors or self.failures or self.unexpectedSuccesses)
def _err_details_to_string(self, test, err=None, details=None):
"""Convert an error in exc_info form or a contents dict to a string."""
if err is not None:
return TracebackContent(
err, test, capture_locals=self.tb_locals).as_text()
return _details_to_str(details, special='traceback')
def _exc_info_to_unicode(self, err, test):
# Deprecated. Only present because subunit upcalls to it. See
# <https://bugs.launchpad.net/testtools/+bug/929063>.
return TracebackContent(err, test).as_text()
def _now(self):
"""Return the current 'test time'.
If the time() method has not been called, this is equivalent to
datetime.now(), otherwise its the last supplied datestamp given to the
time() method.
"""
if self.__now is None:
return datetime.datetime.now(utc)
else:
return self.__now
def startTestRun(self):
"""Called before a test run starts.
New in Python 2.7. The testtools version resets the result to a
pristine condition ready for use in another test run. Note that this
is different from Python 2.7's startTestRun, which does nothing.
"""
# failfast and tb_locals are reset by the super __init__, so save them.
failfast = self.failfast
tb_locals = self.tb_locals
super(TestResult, self).__init__()
self.skip_reasons = {}
self.__now = None
self._tags = TagContext()
# -- Start: As per python 2.7 --
self.expectedFailures = []
self.unexpectedSuccesses = []
self.failfast = failfast
# -- End: As per python 2.7 --
# -- Python 3.5
self.tb_locals = tb_locals
def stopTestRun(self):
"""Called after a test run completes
New in python 2.7
"""
def startTest(self, test):
super(TestResult, self).startTest(test)
self._tags = TagContext(self._tags)
def stopTest(self, test):
self._tags = self._tags.parent
super(TestResult, self).stopTest(test)
@property
def current_tags(self):
"""The currently set tags."""
return self._tags.get_current_tags()
def tags(self, new_tags, gone_tags):
"""Add and remove tags from the test.
:param new_tags: A set of tags to be added to the stream.
:param gone_tags: A set of tags to be removed from the stream.
"""
self._tags.change_tags(new_tags, gone_tags)
def time(self, a_datetime):
"""Provide a timestamp to represent the current time.
This is useful when test activity is time delayed, or happening
concurrently and getting the system time between API calls will not
accurately represent the duration of tests (or the whole run).
Calling time() sets the datetime used by the TestResult object.
Time is permitted to go backwards when using this call.
:param a_datetime: A datetime.datetime object with TZ information or
None to reset the TestResult to gathering time from the system.
"""
self.__now = a_datetime
def done(self):
"""Called when the test runner is done.
deprecated in favour of stopTestRun.
"""
"""Interim states:
* None - no particular status is being reported, or status being reported is
not associated with a test (e.g. when reporting on stdout / stderr chatter).
* inprogress - the test is currently running. Emitted by tests when they start
running and at any intermediary point they might choose to indicate their
continual operation.
"""
INTERIM_STATES = frozenset([None, 'inprogress'])
"""Final states:
* exists - the test exists. This is used when a test is not being executed.
Typically this is when querying what tests could be run in a test run (which
is useful for selecting tests to run).
* xfail - the test failed but that was expected. This is purely informative -
the test is not considered to be a failure.
* uxsuccess - the test passed but was expected to fail. The test will be
considered a failure.
* success - the test has finished without error.
* fail - the test failed (or errored). The test will be considered a failure.
* skip - the test was selected to run but chose to be skipped. e.g. a test
dependency was missing. This is purely informative: the test is not
considered to be a failure.
* unknown - we don't know what state the test is in
"""
FINAL_STATES = frozenset(
['exists', 'xfail', 'uxsuccess', 'success', 'fail', 'skip', 'unknown'])
STATES = INTERIM_STATES | FINAL_STATES
class StreamResult(object):
"""A test result for reporting the activity of a test run.
Typical use
>>> result = StreamResult()
>>> result.startTestRun()
>>> try:
... case.run(result)
... finally:
... result.stopTestRun()
The case object will be either a TestCase or a TestSuite, and
generally make a sequence of calls like::
>>> result.status(self.id(), 'inprogress')
>>> result.status(self.id(), 'success')
General concepts
StreamResult is built to process events that are emitted by tests during a
test run or test enumeration. The test run may be running concurrently, and
even be spread out across multiple machines.
All events are timestamped to prevent network buffering or scheduling
latency causing false timing reports. Timestamps are datetime objects in
the UTC timezone.
A route_code is a unicode string that identifies where a particular test
run. This is optional in the API but very useful when multiplexing multiple
streams together as it allows identification of interactions between tests
that were run on the same hardware or in the same test process. Generally
actual tests never need to bother with this - it is added and processed
by StreamResult's that do multiplexing / run analysis. route_codes are
also used to route stdin back to pdb instances.
The StreamResult base class does no accounting or processing, rather it
just provides an empty implementation of every method, suitable for use
as a base class regardless of intent.
"""
def startTestRun(self):
"""Start a test run.
This will prepare the test result to process results (which might imply
connecting to a database or remote machine).
"""
def stopTestRun(self):
"""Stop a test run.
This informs the result that no more test updates will be received. At
this point any test ids that have started and not completed can be
considered failed-or-hung.
"""
def status(self, test_id=None, test_status=None, test_tags=None,
runnable=True, file_name=None, file_bytes=None, eof=False,
mime_type=None, route_code=None, timestamp=None):
"""Inform the result about a test status.
:param test_id: The test whose status is being reported. None to
report status about the test run as a whole.
:param test_status: The status for the test. There are two sorts of
status - interim and final status events. As many interim events
can be generated as desired, but only one final event. After a
final status event any further file or status events from the
same test_id+route_code may be discarded or associated with a new
test by the StreamResult. (But no exception will be thrown).
Interim states:
* None - no particular status is being reported, or status being
reported is not associated with a test (e.g. when reporting on
stdout / stderr chatter).
* inprogress - the test is currently running. Emitted by tests
when they start running and at any intermediary point they
might choose to indicate their continual operation.
Final states:
* exists - the test exists. This is used when a test is not being
executed. Typically this is when querying what tests could be
run in a test run (which is useful for selecting tests to run).
* xfail - the test failed but that was expected. This is purely
informative - the test is not considered to be a failure.
* uxsuccess - the test passed but was expected to fail. The test
will be considered a failure.
* success - the test has finished without error.
* fail - the test failed (or errored). The test will be
considered a failure.
* skip - the test was selected to run but chose to be skipped.
e.g. a test dependency was missing. This is purely informative-
the test is not considered to be a failure.
:param test_tags: Optional set of tags to apply to the test. Tags
have no intrinsic meaning - that is up to the test author.
:param runnable: Allows status reports to mark that they are for
tests which are not able to be explicitly run. For instance,
subtests will report themselves as non-runnable.
:param file_name: The name for the file_bytes. Any unicode string may
be used. While there is no semantic value attached to the name
of any attachment, the names 'stdout' and 'stderr' and 'traceback'
are recommended for use only for output sent to stdout, stderr and
tracebacks of exceptions. When file_name is supplied, file_bytes
must be a bytes instance.
:param file_bytes: A bytes object containing content for the named
file. This can just be a single chunk of the file - emitting
another file event with more later. Must be None unleses a
file_name is supplied.
:param eof: True if this chunk is the last chunk of the file, any
additional chunks with the same name should be treated as an error
and discarded. Ignored unless file_name has been supplied.
:param mime_type: An optional MIME type for the file. stdout and
stderr will generally be "text/plain; charset=utf8". If None,
defaults to application/octet-stream. Ignored unless file_name
has been supplied.
"""
def domap(function, *sequences):
"""A strict version of 'map' that's guaranteed to run on all inputs.
DEPRECATED since testtools 1.8.1: Internal code should use _strict_map.
External code should look for other solutions for their strict mapping
needs.
"""
warnings.warn(
"domap deprecated since 1.8.1. Please implement your own strict map.",
DeprecationWarning, stacklevel=2)
return _strict_map(function, *sequences)
def _strict_map(function, *sequences):
return list(map(function, *sequences))
class CopyStreamResult(StreamResult):
"""Copies all event it receives to multiple results.
This provides an easy facility for combining multiple StreamResults.
For TestResult the equivalent class was ``MultiTestResult``.
"""
def __init__(self, targets):
super(CopyStreamResult, self).__init__()
self.targets = targets
def startTestRun(self):
super(CopyStreamResult, self).startTestRun()
_strict_map(methodcaller('startTestRun'), self.targets)
def stopTestRun(self):
super(CopyStreamResult, self).stopTestRun()
_strict_map(methodcaller('stopTestRun'), self.targets)
def status(self, *args, **kwargs):
super(CopyStreamResult, self).status(*args, **kwargs)
_strict_map(methodcaller('status', *args, **kwargs), self.targets)
class StreamFailFast(StreamResult):
"""Call the supplied callback if an error is seen in a stream.
An example callback::
def do_something():
pass
"""
def __init__(self, on_error):
self.on_error = on_error
def status(self, test_id=None, test_status=None, test_tags=None,
runnable=True, file_name=None, file_bytes=None, eof=False,
mime_type=None, route_code=None, timestamp=None):
if test_status in ('uxsuccess', 'fail'):
self.on_error()
class StreamResultRouter(StreamResult):
"""A StreamResult that routes events.
StreamResultRouter forwards received events to another StreamResult object,
selected by a dynamic forwarding policy. Events where no destination is
found are forwarded to the fallback StreamResult, or an error is raised.
Typical use is to construct a router with a fallback and then either
create up front mapping rules, or create them as-needed from the fallback
handler::
>>> router = StreamResultRouter()
>>> sink = doubles.StreamResult()
>>> router.add_rule(sink, 'route_code_prefix', route_prefix='0',
... consume_route=True)
>>> router.status(
... test_id='foo', route_code='0/1', test_status='uxsuccess')
StreamResultRouter has no buffering.
When adding routes (and for the fallback) whether to call startTestRun and
stopTestRun or to not call them is controllable by passing
'do_start_stop_run'. The default is to call them for the fallback only.
If a route is added after startTestRun has been called, and
do_start_stop_run is True then startTestRun is called immediately on the
new route sink.
There is no a-priori defined lookup order for routes: if they are ambiguous
the behaviour is undefined. Only a single route is chosen for any event.
"""
_policies = {}
def __init__(self, fallback=None, do_start_stop_run=True):
"""Construct a StreamResultRouter with optional fallback.
:param fallback: A StreamResult to forward events to when no route
exists for them.
:param do_start_stop_run: If False do not pass startTestRun and
stopTestRun onto the fallback.
"""
self.fallback = fallback
self._route_code_prefixes = {}
self._test_ids = {}
# Records sinks that should have do_start_stop_run called on them.
self._sinks = []
if do_start_stop_run and fallback:
self._sinks.append(fallback)
self._in_run = False
def startTestRun(self):
super(StreamResultRouter, self).startTestRun()
for sink in self._sinks:
sink.startTestRun()
self._in_run = True
def stopTestRun(self):
super(StreamResultRouter, self).stopTestRun()
for sink in self._sinks:
sink.stopTestRun()
self._in_run = False
def status(self, **kwargs):
route_code = kwargs.get('route_code', None)
test_id = kwargs.get('test_id', None)
if route_code is not None:
prefix = route_code.split('/')[0]
else:
prefix = route_code
if prefix in self._route_code_prefixes:
target, consume_route = self._route_code_prefixes[prefix]
if route_code is not None and consume_route:
route_code = route_code[len(prefix) + 1:]
if not route_code:
route_code = None
kwargs['route_code'] = route_code
elif test_id in self._test_ids:
target = self._test_ids[test_id]
else:
target = self.fallback
target.status(**kwargs)
def add_rule(self, sink, policy, do_start_stop_run=False, **policy_args):
"""Add a rule to route events to sink when they match a given policy.
:param sink: A StreamResult to receive events.
:param policy: A routing policy. Valid policies are
'route_code_prefix' and 'test_id'.
:param do_start_stop_run: If True then startTestRun and stopTestRun
events will be passed onto this sink.
:raises: ValueError if the policy is unknown
:raises: TypeError if the policy is given arguments it cannot handle.
``route_code_prefix`` routes events based on a prefix of the route
code in the event. It takes a ``route_prefix`` argument to match on
(e.g. '0') and a ``consume_route`` argument, which, if True, removes
the prefix from the ``route_code`` when forwarding events.
``test_id`` routes events based on the test id. It takes a single
argument, ``test_id``. Use ``None`` to select non-test events.
"""
policy_method = StreamResultRouter._policies.get(policy, None)
if not policy_method:
raise ValueError("bad policy %r" % (policy,))
policy_method(self, sink, **policy_args)
if do_start_stop_run:
self._sinks.append(sink)
if self._in_run:
sink.startTestRun()
def _map_route_code_prefix(self, sink, route_prefix, consume_route=False):
if '/' in route_prefix:
raise TypeError(
"%r is more than one route step long" % (route_prefix,))
self._route_code_prefixes[route_prefix] = (sink, consume_route)
_policies['route_code_prefix'] = _map_route_code_prefix
def _map_test_id(self, sink, test_id):
self._test_ids[test_id] = sink
_policies['test_id'] = _map_test_id
class StreamTagger(CopyStreamResult):
"""Adds or discards tags from StreamResult events."""
def __init__(self, targets, add=None, discard=None):
"""Create a StreamTagger.
:param targets: A list of targets to forward events onto.
:param add: Either None or an iterable of tags to add to each event.
:param discard: Either None or an iterable of tags to discard from each
event.
"""
super(StreamTagger, self).__init__(targets)
self.add = frozenset(add or ())
self.discard = frozenset(discard or ())
def status(self, *args, **kwargs):
test_tags = kwargs.get('test_tags') or set()
test_tags.update(self.add)
test_tags.difference_update(self.discard)
kwargs['test_tags'] = test_tags or None
super(StreamTagger, self).status(*args, **kwargs)
class _TestRecord(object):
"""Representation of a test."""
def __init__(self, id, tags, details, status, timestamps):
# The test id.
self.id = id
# Tags for the test.
self.tags = tags
# File attachments.
self.details = details
# One of the StreamResult status codes.
self.status = status
# Pair of timestamps (x, y).
# x is the first timestamp we received for this test, y is the one that
# triggered the notification. y can be None if the test hanged.
self.timestamps = timestamps
@classmethod
def create(cls, test_id, timestamp):
return cls(
id=test_id,
tags=set(),
details={},
status='unknown',
timestamps=(timestamp, None),
)
def set(self, *args, **kwargs):
if args:
setattr(self, args[0], args[1])
for key, value in kwargs.items():
setattr(self, key, value)
return self
def transform(self, data, value):
getattr(self, data[0])[data[1]] = value
return self
def to_dict(self):
"""Convert record into a "test dict".
A "test dict" is a concept used in other parts of the code-base. It
has the following keys:
* id: the test id.
* tags: The tags for the test. A set of unicode strings.
* details: A dict of file attachments - ``testtools.content.Content``
objects.
* status: One of the StreamResult status codes (including inprogress)
or 'unknown' (used if only file events for a test were received...)
* timestamps: A pair of timestamps - the first one received with this
test id, and the one in the event that triggered the notification.
Hung tests have a None for the second end event. Timestamps are not
compared - their ordering is purely order received in the stream.
"""
return {
'id': self.id,
'tags': self.tags,
'details': self.details,
'status': self.status,
'timestamps': list(self.timestamps),
}
def got_timestamp(self, timestamp):
"""Called when we receive a timestamp.
This will always update the second element of the 'timestamps' tuple.
It doesn't compare timestamps at all.
"""
return self.set(timestamps=(self.timestamps[0], timestamp))
def got_file(self, file_name, file_bytes, mime_type=None):
"""Called when we receive file information.
``mime_type`` is only used when this is the first time we've seen data
from this file.
"""
if file_name in self.details:
case = self
else:
content_type = _make_content_type(mime_type)
content_bytes = []
case = self.transform(
['details', file_name],
Content(content_type, lambda: content_bytes))
case.details[file_name].iter_bytes().append(file_bytes)
return case
def to_test_case(self):
"""Convert into a TestCase object.
:return: A PlaceHolder test object.
"""
# Circular import.
global PlaceHolder
if PlaceHolder is None:
from testtools.testcase import PlaceHolder
outcome = _status_map[self.status]
return PlaceHolder(
self.id,
outcome=outcome,
details=self.details,
tags=self.tags,
timestamps=self.timestamps,
)
def _make_content_type(mime_type=None):
"""Return ContentType for a given mime type.
testtools was emitting a bad encoding, and this works around it.
Unfortunately, is also loses data - probably want to drop this in a few
releases.
"""
# XXX: Not sure what release this was added, so "in a few releases" is
# unactionable.
if mime_type is None:
mime_type = 'application/octet-stream'
primary, sub, parameters = parse_mime_type(mime_type)
if 'charset' in parameters:
if ',' in parameters['charset']:
parameters['charset'] = parameters['charset'][
:parameters['charset'].find(',')]
return ContentType(primary, sub, parameters)
_status_map = {
'inprogress': 'addFailure',
'unknown': 'addFailure',
'success': 'addSuccess',
'skip': 'addSkip',
'fail': 'addFailure',
'xfail': 'addExpectedFailure',
'uxsuccess': 'addUnexpectedSuccess',
}
class _StreamToTestRecord(StreamResult):
"""A specialised StreamResult that emits a callback as tests complete.
Top level file attachments are simply discarded. Hung tests are detected
by stopTestRun and notified there and then.
The callback is passed a ``_TestRecord`` object.
Only the most recent tags observed in the stream are reported.
"""
def __init__(self, on_test):
"""Create a _StreamToTestRecord calling on_test on test completions.
:param on_test: A callback that accepts one parameter:
a ``_TestRecord`` object describing a test.
"""
super(_StreamToTestRecord, self).__init__()
self.on_test = on_test
if parse_mime_type is None:
raise ImportError("mimeparse module missing.")
def startTestRun(self):
super(_StreamToTestRecord, self).startTestRun()
self._inprogress = {}
def status(self, test_id=None, test_status=None, test_tags=None,
runnable=True, file_name=None, file_bytes=None, eof=False,
mime_type=None, route_code=None, timestamp=None):
super(_StreamToTestRecord, self).status(
test_id, test_status,
test_tags=test_tags, runnable=runnable, file_name=file_name,
file_bytes=file_bytes, eof=eof, mime_type=mime_type,
route_code=route_code, timestamp=timestamp)
key = self._ensure_key(test_id, route_code, timestamp)
if not key:
return
# update fields
self._inprogress[key] = self._update_case(
self._inprogress[key], test_status, test_tags, file_name,
file_bytes, mime_type, timestamp)
# notify completed tests.
if test_status not in INTERIM_STATES:
self.on_test(self._inprogress.pop(key))
def _update_case(self, case, test_status=None, test_tags=None,
file_name=None, file_bytes=None, mime_type=None,
timestamp=None):
if test_status is not None:
case = case.set(status=test_status)
case = case.got_timestamp(timestamp)
if file_name is not None and file_bytes:
case = case.got_file(file_name, file_bytes, mime_type)
if test_tags is not None:
case = case.set('tags', test_tags)
return case
def stopTestRun(self):
super(_StreamToTestRecord, self).stopTestRun()
while self._inprogress:
case = self._inprogress.popitem()[1]
self.on_test(case.got_timestamp(None))
def _ensure_key(self, test_id, route_code, timestamp):
if test_id is None:
return
key = (test_id, route_code)
if key not in self._inprogress:
self._inprogress[key] = _TestRecord.create(test_id, timestamp)
return key
class StreamToDict(StreamResult):
"""A specialised StreamResult that emits a callback as tests complete.
Top level file attachments are simply discarded. Hung tests are detected
by stopTestRun and notified there and then.
The callback is passed a dict with the following keys:
* id: the test id.
* tags: The tags for the test. A set of unicode strings.
* details: A dict of file attachments - ``testtools.content.Content``
objects.
* status: One of the StreamResult status codes (including inprogress) or
'unknown' (used if only file events for a test were received...)
* timestamps: A pair of timestamps - the first one received with this
test id, and the one in the event that triggered the notification.
Hung tests have a None for the second end event. Timestamps are not
compared - their ordering is purely order received in the stream.
Only the most recent tags observed in the stream are reported.
"""
# XXX: This could actually be replaced by a very simple function.
# Unfortunately, subclassing is a supported API.
# XXX: Alternative simplification is to extract a StreamAdapter base
# class, and have this inherit from that.
def __init__(self, on_test):
"""Create a _StreamToTestRecord calling on_test on test completions.
:param on_test: A callback that accepts one parameter:
a dictionary describing a test.
"""
super(StreamToDict, self).__init__()
self._hook = _StreamToTestRecord(self._handle_test)
# XXX: Not clear whether its part of the supported interface for
# self.on_test to be the passed-in on_test. If not, we could reduce
# the boilerplate by subclassing _StreamToTestRecord.
self.on_test = on_test
def _handle_test(self, test_record):
self.on_test(test_record.to_dict())
def startTestRun(self):
super(StreamToDict, self).startTestRun()
self._hook.startTestRun()
def status(self, *args, **kwargs):
super(StreamToDict, self).status(*args, **kwargs)
self._hook.status(*args, **kwargs)
def stopTestRun(self):
super(StreamToDict, self).stopTestRun()
self._hook.stopTestRun()
def test_dict_to_case(test_dict):
"""Convert a test dict into a TestCase object.
:param test_dict: A test dict as generated by StreamToDict.
:return: A PlaceHolder test object.
"""
return _TestRecord(
id=test_dict['id'],
tags=test_dict['tags'],
details=test_dict['details'],
status=test_dict['status'],
timestamps=tuple(test_dict['timestamps']),
).to_test_case()
class StreamSummary(StreamResult):
"""A specialised StreamResult that summarises a stream.
The summary uses the same representation as the original
unittest.TestResult contract, allowing it to be consumed by any test
runner.
"""
def __init__(self):
super(StreamSummary, self).__init__()
self._hook = _StreamToTestRecord(self._gather_test)
self._handle_status = {
'success': self._success,
'skip': self._skip,
'exists': self._exists,
'fail': self._fail,
'xfail': self._xfail,
'uxsuccess': self._uxsuccess,
'unknown': self._incomplete,
'inprogress': self._incomplete,
}
def startTestRun(self):
super(StreamSummary, self).startTestRun()
self.failures = []
self.errors = []
self.testsRun = 0
self.skipped = []
self.expectedFailures = []
self.unexpectedSuccesses = []
self._hook.startTestRun()
def status(self, *args, **kwargs):
super(StreamSummary, self).status(*args, **kwargs)
self._hook.status(*args, **kwargs)
def stopTestRun(self):
super(StreamSummary, self).stopTestRun()
self._hook.stopTestRun()
def wasSuccessful(self):
"""Return False if any failure has occured.
Note that incomplete tests can only be detected when stopTestRun is
called, so that should be called before checking wasSuccessful.
"""
return (not self.failures and not self.errors)
def _gather_test(self, test_record):
if test_record.status == 'exists':
return
self.testsRun += 1
case = test_record.to_test_case()
self._handle_status[test_record.status](case)
def _incomplete(self, case):
self.errors.append((case, "Test did not complete"))
def _success(self, case):
pass
def _skip(self, case):
if 'reason' not in case._details:
reason = "Unknown"
else:
reason = case._details['reason'].as_text()
self.skipped.append((case, reason))
def _exists(self, case):
pass
def _fail(self, case):
message = _details_to_str(case._details, special="traceback")
self.errors.append((case, message))
def _xfail(self, case):
message = _details_to_str(case._details, special="traceback")
self.expectedFailures.append((case, message))
def _uxsuccess(self, case):
case._outcome = 'addUnexpectedSuccess'
self.unexpectedSuccesses.append(case)
class TestControl(object):
"""Controls a running test run, allowing it to be interrupted.
:ivar shouldStop: If True, tests should not run and should instead
return immediately. Similarly a TestSuite should check this between
each test and if set stop dispatching any new tests and return.
"""
def __init__(self):
super(TestControl, self).__init__()
self.shouldStop = False
def stop(self):
"""Indicate that tests should stop running."""
self.shouldStop = True
class MultiTestResult(TestResult):
"""A test result that dispatches to many test results."""
def __init__(self, *results):
# Setup _results first, as the base class __init__ assigns to failfast.
self._results = list(map(ExtendedToOriginalDecorator, results))
super(MultiTestResult, self).__init__()
def __repr__(self):
return '<%s (%s)>' % (
self.__class__.__name__, ', '.join(map(repr, self._results)))
def _dispatch(self, message, *args, **kwargs):
return tuple(
getattr(result, message)(*args, **kwargs)
for result in self._results)
def _get_failfast(self):
return getattr(self._results[0], 'failfast', False)
def _set_failfast(self, value):
self._dispatch('__setattr__', 'failfast', value)
failfast = property(_get_failfast, _set_failfast)
def _get_shouldStop(self):
return any(self._dispatch('__getattr__', 'shouldStop'))
def _set_shouldStop(self, value):
# Called because we subclass TestResult. Probably should not do that.
pass
shouldStop = property(_get_shouldStop, _set_shouldStop)
def startTest(self, test):
super(MultiTestResult, self).startTest(test)
return self._dispatch('startTest', test)
def stop(self):
return self._dispatch('stop')
def stopTest(self, test):
super(MultiTestResult, self).stopTest(test)
return self._dispatch('stopTest', test)
def addError(self, test, error=None, details=None):
return self._dispatch('addError', test, error, details=details)
def addExpectedFailure(self, test, err=None, details=None):
return self._dispatch(
'addExpectedFailure', test, err, details=details)
def addFailure(self, test, err=None, details=None):
return self._dispatch('addFailure', test, err, details=details)
def addSkip(self, test, reason=None, details=None):
return self._dispatch('addSkip', test, reason, details=details)
def addSuccess(self, test, details=None):
return self._dispatch('addSuccess', test, details=details)
def addUnexpectedSuccess(self, test, details=None):
return self._dispatch('addUnexpectedSuccess', test, details=details)
def startTestRun(self):
super(MultiTestResult, self).startTestRun()
return self._dispatch('startTestRun')
def stopTestRun(self):
return self._dispatch('stopTestRun')
def tags(self, new_tags, gone_tags):
super(MultiTestResult, self).tags(new_tags, gone_tags)
return self._dispatch('tags', new_tags, gone_tags)
def time(self, a_datetime):
return self._dispatch('time', a_datetime)
def done(self):
return self._dispatch('done')
def wasSuccessful(self):
"""Was this result successful?
Only returns True if every constituent result was successful.
"""
return all(self._dispatch('wasSuccessful'))
class TextTestResult(TestResult):
"""A TestResult which outputs activity to a text stream."""
def __init__(self, stream, failfast=False, tb_locals=False):
"""Construct a TextTestResult writing to stream."""
super(TextTestResult, self).__init__(
failfast=failfast, tb_locals=tb_locals)
self.stream = stream
self.sep1 = '=' * 70 + '\n'
self.sep2 = '-' * 70 + '\n'
def _delta_to_float(self, a_timedelta, precision):
# This calls ceiling to ensure that the most pessimistic view of time
# taken is shown (rather than leaving it to the Python %f operator
# to decide whether to round/floor/ceiling. This was added when we
# had pyp3 test failures that suggest a floor was happening.
shift = 10 ** precision
return math.ceil(
(a_timedelta.days * 86400.0 + a_timedelta.seconds +
a_timedelta.microseconds / 1000000.0) * shift) / shift
def _show_list(self, label, error_list):
for test, output in error_list:
self.stream.write(self.sep1)
self.stream.write("%s: %s\n" % (label, test.id()))
self.stream.write(self.sep2)
self.stream.write(output)
def startTestRun(self):
super(TextTestResult, self).startTestRun()
self.__start = self._now()
self.stream.write("Tests running...\n")
def stopTestRun(self):
if self.testsRun != 1:
plural = 's'
else:
plural = ''
stop = self._now()
self._show_list('ERROR', self.errors)
self._show_list('FAIL', self.failures)
for test in self.unexpectedSuccesses:
self.stream.write(
"%sUNEXPECTED SUCCESS: %s\n%s" % (
self.sep1, test.id(), self.sep2))
self.stream.write(
"\nRan %d test%s in %.3fs\n" % (
self.testsRun, plural,
self._delta_to_float(stop - self.__start, 3)))
if self.wasSuccessful():
self.stream.write("OK\n")
else:
self.stream.write("FAILED (")
details = []
details.append("failures=%d" % (
sum(map(len, (
self.failures, self.errors, self.unexpectedSuccesses)))))
self.stream.write(", ".join(details))
self.stream.write(")\n")
super(TextTestResult, self).stopTestRun()
class ThreadsafeForwardingResult(TestResult):
"""A TestResult which ensures the target does not receive mixed up calls.
Multiple ``ThreadsafeForwardingResults`` can forward to the same target
result, and that target result will only ever receive the complete set of
events for one test at a time.
This is enforced using a semaphore, which further guarantees that tests
will be sent atomically even if the ``ThreadsafeForwardingResults`` are in
different threads.
``ThreadsafeForwardingResult`` is typically used by
``ConcurrentTestSuite``, which creates one ``ThreadsafeForwardingResult``
per thread, each of which wraps of the TestResult that
``ConcurrentTestSuite.run()`` is called with.
target.startTestRun() and target.stopTestRun() are called once for each
ThreadsafeForwardingResult that forwards to the same target. If the target
takes special action on these events, it should take care to accommodate
this.
time() and tags() calls are batched to be adjacent to the test result and
in the case of tags() are coerced into test-local scope, avoiding the
opportunity for bugs around global state in the target.
"""
def __init__(self, target, semaphore):
"""Create a ThreadsafeForwardingResult forwarding to target.
:param target: A ``TestResult``.
:param semaphore: A ``threading.Semaphore`` with limit 1.
"""
TestResult.__init__(self)
self.result = ExtendedToOriginalDecorator(target)
self.semaphore = semaphore
self._test_start = None
self._global_tags = set(), set()
self._test_tags = set(), set()
def __repr__(self):
return '<%s %r>' % (self.__class__.__name__, self.result)
def _any_tags(self, tags):
return bool(tags[0] or tags[1])
def _add_result_with_semaphore(self, method, test, *args, **kwargs):
now = self._now()
self.semaphore.acquire()
try:
self.result.time(self._test_start)
self.result.startTest(test)
self.result.time(now)
if self._any_tags(self._global_tags):
self.result.tags(*self._global_tags)
if self._any_tags(self._test_tags):
self.result.tags(*self._test_tags)
self._test_tags = set(), set()
try:
method(test, *args, **kwargs)
finally:
self.result.stopTest(test)
finally:
self.semaphore.release()
self._test_start = None
def addError(self, test, err=None, details=None):
self._add_result_with_semaphore(
self.result.addError, test, err, details=details)
def addExpectedFailure(self, test, err=None, details=None):
self._add_result_with_semaphore(
self.result.addExpectedFailure, test, err, details=details)
def addFailure(self, test, err=None, details=None):
self._add_result_with_semaphore(
self.result.addFailure, test, err, details=details)
def addSkip(self, test, reason=None, details=None):
self._add_result_with_semaphore(
self.result.addSkip, test, reason, details=details)
def addSuccess(self, test, details=None):
self._add_result_with_semaphore(
self.result.addSuccess, test, details=details)
def addUnexpectedSuccess(self, test, details=None):
self._add_result_with_semaphore(
self.result.addUnexpectedSuccess, test, details=details)
def progress(self, offset, whence):
pass
def startTestRun(self):
super(ThreadsafeForwardingResult, self).startTestRun()
self.semaphore.acquire()
try:
self.result.startTestRun()
finally:
self.semaphore.release()
def _get_shouldStop(self):
self.semaphore.acquire()
try:
return self.result.shouldStop
finally:
self.semaphore.release()
def _set_shouldStop(self, value):
# Another case where we should not subclass TestResult
pass
shouldStop = property(_get_shouldStop, _set_shouldStop)
def stop(self):
self.semaphore.acquire()
try:
self.result.stop()
finally:
self.semaphore.release()
def stopTestRun(self):
self.semaphore.acquire()
try:
self.result.stopTestRun()
finally:
self.semaphore.release()
def done(self):
self.semaphore.acquire()
try:
self.result.done()
finally:
self.semaphore.release()
def startTest(self, test):
self._test_start = self._now()
super(ThreadsafeForwardingResult, self).startTest(test)
def wasSuccessful(self):
return self.result.wasSuccessful()
def tags(self, new_tags, gone_tags):
"""See `TestResult`."""
super(ThreadsafeForwardingResult, self).tags(new_tags, gone_tags)
if self._test_start is not None:
self._test_tags = _merge_tags(
self._test_tags, (new_tags, gone_tags))
else:
self._global_tags = _merge_tags(
self._global_tags, (new_tags, gone_tags))
def _merge_tags(existing, changed):
new_tags, gone_tags = changed
result_new = set(existing[0])
result_gone = set(existing[1])
result_new.update(new_tags)
result_new.difference_update(gone_tags)
result_gone.update(gone_tags)
result_gone.difference_update(new_tags)
return result_new, result_gone
class ExtendedToOriginalDecorator(object):
"""Permit new TestResult API code to degrade gracefully with old results.
This decorates an existing TestResult and converts missing outcomes
such as addSkip to older outcomes such as addSuccess. It also supports
the extended details protocol. In all cases the most recent protocol
is attempted first, and fallbacks only occur when the decorated result
does not support the newer style of calling.
"""
def __init__(self, decorated):
self.decorated = decorated
self._tags = TagContext()
# Only used for old TestResults that do not have failfast.
self._failfast = False
# Used for old TestResults that do not have stop.
self._shouldStop = False
def __repr__(self):
return '<%s %r>' % (self.__class__.__name__, self.decorated)
def __getattr__(self, name):
return getattr(self.decorated, name)
def addError(self, test, err=None, details=None):
try:
self._check_args(err, details)
if details is not None:
try:
return self.decorated.addError(test, details=details)
except TypeError:
# have to convert
err = self._details_to_exc_info(details)
return self.decorated.addError(test, err)
finally:
if self.failfast:
self.stop()
def addExpectedFailure(self, test, err=None, details=None):
self._check_args(err, details)
addExpectedFailure = getattr(
self.decorated, 'addExpectedFailure', None)
if addExpectedFailure is None:
return self.addSuccess(test)
if details is not None:
try:
return addExpectedFailure(test, details=details)
except TypeError:
# have to convert
err = self._details_to_exc_info(details)
return addExpectedFailure(test, err)
def addFailure(self, test, err=None, details=None):
try:
self._check_args(err, details)
if details is not None:
try:
return self.decorated.addFailure(test, details=details)
except TypeError:
# have to convert
err = self._details_to_exc_info(details)
return self.decorated.addFailure(test, err)
finally:
if self.failfast:
self.stop()
def addSkip(self, test, reason=None, details=None):
self._check_args(reason, details)
addSkip = getattr(self.decorated, 'addSkip', None)
if addSkip is None:
return self.decorated.addSuccess(test)
if details is not None:
try:
return addSkip(test, details=details)
except TypeError:
# extract the reason if it's available
try:
reason = details['reason'].as_text()
except KeyError:
reason = _details_to_str(details)
return addSkip(test, reason)
def addUnexpectedSuccess(self, test, details=None):
try:
outcome = getattr(self.decorated, 'addUnexpectedSuccess', None)
if outcome is None:
try:
test.fail("")
except test.failureException:
return self.addFailure(test, sys.exc_info())
if details is not None:
try:
return outcome(test, details=details)
except TypeError:
pass
return outcome(test)
finally:
if self.failfast:
self.stop()
def addSuccess(self, test, details=None):
if details is not None:
try:
return self.decorated.addSuccess(test, details=details)
except TypeError:
pass
return self.decorated.addSuccess(test)
def _check_args(self, err, details):
param_count = 0
if err is not None:
param_count += 1
if details is not None:
param_count += 1
if param_count != 1:
raise ValueError(
"Must pass only one of err '%s' and details '%s"
% (err, details))
def _details_to_exc_info(self, details):
"""Convert a details dict to an exc_info tuple."""
return (
_StringException,
_StringException(_details_to_str(details, special='traceback')),
None)
@property
def current_tags(self):
return getattr(
self.decorated, 'current_tags', self._tags.get_current_tags())
def done(self):
try:
return self.decorated.done()
except AttributeError:
return
def _get_failfast(self):
return getattr(self.decorated, 'failfast', self._failfast)
def _set_failfast(self, value):
if safe_hasattr(self.decorated, 'failfast'):
self.decorated.failfast = value
else:
self._failfast = value
failfast = property(_get_failfast, _set_failfast)
def progress(self, offset, whence):
method = getattr(self.decorated, 'progress', None)
if method is None:
return
return method(offset, whence)
def _get_shouldStop(self):
return getattr(self.decorated, 'shouldStop', self._shouldStop)
def _set_shouldStop(self, value):
if safe_hasattr(self.decorated, 'shouldStop'):
self.decorated.shouldStop = value
else:
self._shouldStop = value
shouldStop = property(_get_shouldStop, _set_shouldStop)
def startTest(self, test):
self._tags = TagContext(self._tags)
return self.decorated.startTest(test)
def startTestRun(self):
self._tags = TagContext()
try:
return self.decorated.startTestRun()
except AttributeError:
return
def stop(self):
method = getattr(self.decorated, 'stop', None)
if method:
return method()
self.shouldStop = True
def stopTest(self, test):
self._tags = self._tags.parent
return self.decorated.stopTest(test)
def stopTestRun(self):
try:
return self.decorated.stopTestRun()
except AttributeError:
return
def tags(self, new_tags, gone_tags):
method = getattr(self.decorated, 'tags', None)
if method is not None:
return method(new_tags, gone_tags)
else:
self._tags.change_tags(new_tags, gone_tags)
def time(self, a_datetime):
method = getattr(self.decorated, 'time', None)
if method is None:
return
return method(a_datetime)
def wasSuccessful(self):
return self.decorated.wasSuccessful()
class ExtendedToStreamDecorator(CopyStreamResult, StreamSummary, TestControl):
"""Permit using old TestResult API code with new StreamResult objects.
This decorates a StreamResult and converts old (Python 2.6 / 2.7 /
Extended) TestResult API calls into StreamResult calls.
It also supports regular StreamResult calls, making it safe to wrap around
any StreamResult.
"""
def __init__(self, decorated):
super(ExtendedToStreamDecorator, self).__init__([decorated])
# Deal with mismatched base class constructors.
TestControl.__init__(self)
self._started = False
def _get_failfast(self):
return len(self.targets) == 2
def _set_failfast(self, value):
if value:
if len(self.targets) == 2:
return
self.targets.append(StreamFailFast(self.stop))
else:
del self.targets[1:]
failfast = property(_get_failfast, _set_failfast)
def startTest(self, test):
if not self._started:
self.startTestRun()
self.status(
test_id=test.id(), test_status='inprogress', timestamp=self._now())
self._tags = TagContext(self._tags)
def stopTest(self, test):
self._tags = self._tags.parent
def addError(self, test, err=None, details=None):
self._check_args(err, details)
self._convert(test, err, details, 'fail')
addFailure = addError
def _convert(self, test, err, details, status, reason=None):
if not self._started:
self.startTestRun()
test_id = test.id()
now = self._now()
if err is not None:
if details is None:
details = {}
details['traceback'] = TracebackContent(err, test)
if details is not None:
for name, content in details.items():
mime_type = repr(content.content_type)
file_bytes = None
for next_bytes in content.iter_bytes():
if file_bytes is not None:
self.status(
file_name=name, file_bytes=file_bytes,
mime_type=mime_type, test_id=test_id,
timestamp=now)
file_bytes = next_bytes
if file_bytes is None:
file_bytes = _b("")
self.status(
file_name=name, file_bytes=file_bytes, eof=True,
mime_type=mime_type, test_id=test_id, timestamp=now)
if reason is not None:
self.status(
file_name='reason', file_bytes=reason.encode('utf8'),
eof=True, mime_type="text/plain; charset=utf8",
test_id=test_id, timestamp=now)
self.status(
test_id=test_id, test_status=status,
test_tags=self.current_tags, timestamp=now)
def addExpectedFailure(self, test, err=None, details=None):
self._check_args(err, details)
self._convert(test, err, details, 'xfail')
def addSkip(self, test, reason=None, details=None):
self._convert(test, None, details, 'skip', reason)
def addUnexpectedSuccess(self, test, details=None):
self._convert(test, None, details, 'uxsuccess')
def addSuccess(self, test, details=None):
self._convert(test, None, details, 'success')
def _check_args(self, err, details):
param_count = 0
if err is not None:
param_count += 1
if details is not None:
param_count += 1
if param_count != 1:
raise ValueError(
"Must pass only one of err '%s' and details '%s"
% (err, details))
def startTestRun(self):
super(ExtendedToStreamDecorator, self).startTestRun()
self._tags = TagContext()
self.shouldStop = False
self.__now = None
self._started = True
@property
def current_tags(self):
"""The currently set tags."""
return self._tags.get_current_tags()
def tags(self, new_tags, gone_tags):
"""Add and remove tags from the test.
:param new_tags: A set of tags to be added to the stream.
:param gone_tags: A set of tags to be removed from the stream.
"""
self._tags.change_tags(new_tags, gone_tags)
def _now(self):
"""Return the current 'test time'.
If the time() method has not been called, this is equivalent to
datetime.now(), otherwise its the last supplied datestamp given to the
time() method.
"""
if self.__now is None:
return datetime.datetime.now(utc)
else:
return self.__now
def time(self, a_datetime):
self.__now = a_datetime
def wasSuccessful(self):
if not self._started:
self.startTestRun()
return super(ExtendedToStreamDecorator, self).wasSuccessful()
class ResourcedToStreamDecorator(ExtendedToStreamDecorator):
"""Report ``testresources``-related activity to StreamResult objects.
Implement the resource lifecycle TestResult protocol extension supported
by the ``testresources.TestResourceManager`` class. At each stage of a
resource's lifecycle, a stream event with relevant details will be
emitted.
Each stream event will have its test_id field set to the resource manager's
identifier (see ``testresources.TestResourceManager.id()``) plus the method
being executed (either 'make' or 'clean').
The test_status will be either 'inprogress' or 'success'.
The runnable flag will be set to False.
"""
def startMakeResource(self, resource):
self._convertResourceLifecycle(resource, 'make', 'start')
def stopMakeResource(self, resource):
self._convertResourceLifecycle(resource, 'make', 'stop')
def startCleanResource(self, resource):
self._convertResourceLifecycle(resource, 'clean', 'start')
def stopCleanResource(self, resource):
self._convertResourceLifecycle(resource, 'clean', 'stop')
def _convertResourceLifecycle(self, resource, method, phase):
"""Convert a resource lifecycle report to a stream event."""
# If the resource implements the TestResourceManager.id() API, let's
# use it, otherwise fallback to the class name.
if safe_hasattr(resource, "id"):
resource_id = resource.id()
else:
resource_id = "%s.%s" % (
resource.__class__.__module__, resource.__class__.__name__)
test_id = '%s.%s' % (resource_id, method)
if phase == 'start':
test_status = 'inprogress'
else:
test_status = 'success'
self.status(
test_id=test_id, test_status=test_status, runnable=False,
timestamp=self._now())
class StreamToExtendedDecorator(StreamResult):
"""Convert StreamResult API calls into ExtendedTestResult calls.
This will buffer all calls for all concurrently active tests, and
then flush each test as they complete.
Incomplete tests will be flushed as errors when the test run stops.
Non test file attachments are accumulated into a test called
'testtools.extradata' flushed at the end of the run.
"""
def __init__(self, decorated):
# ExtendedToOriginalDecorator takes care of thunking details back to
# exceptions/reasons etc.
self.decorated = ExtendedToOriginalDecorator(decorated)
# _StreamToTestRecord buffers and gives us individual tests.
self.hook = _StreamToTestRecord(self._handle_tests)
def status(self, test_id=None, test_status=None, *args, **kwargs):
if test_status == 'exists':
return
self.hook.status(
test_id=test_id, test_status=test_status, *args, **kwargs)
def startTestRun(self):
self.decorated.startTestRun()
self.hook.startTestRun()
def stopTestRun(self):
self.hook.stopTestRun()
self.decorated.stopTestRun()
def _handle_tests(self, test_record):
case = test_record.to_test_case()
case.run(self.decorated)
class StreamToQueue(StreamResult):
"""A StreamResult which enqueues events as a dict to a queue.Queue.
Events have their route code updated to include the route code
StreamToQueue was constructed with before they are submitted. If the event
route code is None, it is replaced with the StreamToQueue route code,
otherwise it is prefixed with the supplied code + a hyphen.
startTestRun and stopTestRun are forwarded to the queue. Implementors that
dequeue events back into StreamResult calls should take care not to call
startTestRun / stopTestRun on other StreamResult objects multiple times
(e.g. by filtering startTestRun and stopTestRun).
``StreamToQueue`` is typically used by
``ConcurrentStreamTestSuite``, which creates one ``StreamToQueue``
per thread, forwards status events to the the StreamResult that
``ConcurrentStreamTestSuite.run()`` was called with, and uses the
stopTestRun event to trigger calling join() on the each thread.
Unlike ThreadsafeForwardingResult which this supercedes, no buffering takes
place - any event supplied to a StreamToQueue will be inserted into the
queue immediately.
Events are forwarded as a dict with a key ``event`` which is one of
``startTestRun``, ``stopTestRun`` or ``status``. When ``event`` is
``status`` the dict also has keys matching the keyword arguments
of ``StreamResult.status``, otherwise it has one other key ``result`` which
is the result that invoked ``startTestRun``.
"""
def __init__(self, queue, routing_code):
"""Create a StreamToQueue forwarding to target.
:param queue: A ``queue.Queue`` to receive events.
:param routing_code: The routing code to apply to messages.
"""
super(StreamToQueue, self).__init__()
self.queue = queue
self.routing_code = routing_code
def startTestRun(self):
self.queue.put(dict(event='startTestRun', result=self))
def status(self, test_id=None, test_status=None, test_tags=None,
runnable=True, file_name=None, file_bytes=None, eof=False,
mime_type=None, route_code=None, timestamp=None):
self.queue.put(dict(
event='status', test_id=test_id,
test_status=test_status, test_tags=test_tags, runnable=runnable,
file_name=file_name, file_bytes=file_bytes, eof=eof,
mime_type=mime_type, route_code=self.route_code(route_code),
timestamp=timestamp))
def stopTestRun(self):
self.queue.put(dict(event='stopTestRun', result=self))
def route_code(self, route_code):
"""Adjust route_code on the way through."""
if route_code is None:
return self.routing_code
return self.routing_code + _u("/") + route_code
class TestResultDecorator(object):
"""General pass-through decorator.
This provides a base that other TestResults can inherit from to
gain basic forwarding functionality.
"""
def __init__(self, decorated):
"""Create a TestResultDecorator forwarding to decorated."""
self.decorated = decorated
def startTest(self, test):
return self.decorated.startTest(test)
def startTestRun(self):
return self.decorated.startTestRun()
def stopTest(self, test):
return self.decorated.stopTest(test)
def stopTestRun(self):
return self.decorated.stopTestRun()
def addError(self, test, err=None, details=None):
return self.decorated.addError(test, err, details=details)
def addFailure(self, test, err=None, details=None):
return self.decorated.addFailure(test, err, details=details)
def addSuccess(self, test, details=None):
return self.decorated.addSuccess(test, details=details)
def addSkip(self, test, reason=None, details=None):
return self.decorated.addSkip(test, reason, details=details)
def addExpectedFailure(self, test, err=None, details=None):
return self.decorated.addExpectedFailure(test, err, details=details)
def addUnexpectedSuccess(self, test, details=None):
return self.decorated.addUnexpectedSuccess(test, details=details)
def progress(self, offset, whence):
return self.decorated.progress(offset, whence)
def wasSuccessful(self):
return self.decorated.wasSuccessful()
@property
def current_tags(self):
return self.decorated.current_tags
@property
def shouldStop(self):
return self.decorated.shouldStop
def stop(self):
return self.decorated.stop()
@property
def testsRun(self):
return self.decorated.testsRun
def tags(self, new_tags, gone_tags):
return self.decorated.tags(new_tags, gone_tags)
def time(self, a_datetime):
return self.decorated.time(a_datetime)
class Tagger(TestResultDecorator):
"""Tag each test individually."""
def __init__(self, decorated, new_tags, gone_tags):
"""Wrap 'decorated' such that each test is tagged.
:param new_tags: Tags to be added for each test.
:param gone_tags: Tags to be removed for each test.
"""
super(Tagger, self).__init__(decorated)
self._new_tags = set(new_tags)
self._gone_tags = set(gone_tags)
def startTest(self, test):
super(Tagger, self).startTest(test)
self.tags(self._new_tags, self._gone_tags)
class TestByTestResult(TestResult):
"""Call something every time a test completes."""
def __init__(self, on_test):
"""Construct a ``TestByTestResult``.
:param on_test: A callable that take a test case, a status (one of
"success", "failure", "error", "skip", or "xfail"), a start time
(a ``datetime`` with timezone), a stop time, an iterable of tags,
and a details dict. Is called at the end of each test (i.e. on
``stopTest``) with the accumulated values for that test.
"""
super(TestByTestResult, self).__init__()
self._on_test = on_test
def startTest(self, test):
super(TestByTestResult, self).startTest(test)
self._start_time = self._now()
# There's no supported (i.e. tested) behaviour that relies on these
# being set, but it makes me more comfortable all the same. -- jml
self._status = None
self._details = None
self._stop_time = None
def stopTest(self, test):
self._stop_time = self._now()
tags = set(self.current_tags)
super(TestByTestResult, self).stopTest(test)
self._on_test(
test=test,
status=self._status,
start_time=self._start_time,
stop_time=self._stop_time,
tags=tags,
details=self._details)
def _err_to_details(self, test, err, details):
if details:
return details
return {'traceback': TracebackContent(
err, test, capture_locals=self.tb_locals)}
def addSuccess(self, test, details=None):
super(TestByTestResult, self).addSuccess(test)
self._status = 'success'
self._details = details
def addFailure(self, test, err=None, details=None):
super(TestByTestResult, self).addFailure(test, err, details)
self._status = 'failure'
self._details = self._err_to_details(test, err, details)
def addError(self, test, err=None, details=None):
super(TestByTestResult, self).addError(test, err, details)
self._status = 'error'
self._details = self._err_to_details(test, err, details)
def addSkip(self, test, reason=None, details=None):
super(TestByTestResult, self).addSkip(test, reason, details)
self._status = 'skip'
if details is None:
details = {'reason': text_content(reason)}
elif reason:
# XXX: What if details already has 'reason' key?
details['reason'] = text_content(reason)
self._details = details
def addExpectedFailure(self, test, err=None, details=None):
super(TestByTestResult, self).addExpectedFailure(test, err, details)
self._status = 'xfail'
self._details = self._err_to_details(test, err, details)
def addUnexpectedSuccess(self, test, details=None):
super(TestByTestResult, self).addUnexpectedSuccess(test, details)
self._status = 'success'
self._details = details
class TimestampingStreamResult(CopyStreamResult):
"""A StreamResult decorator that assigns a timestamp when none is present.
This is convenient for ensuring events are timestamped.
"""
def __init__(self, target):
super(TimestampingStreamResult, self).__init__([target])
def status(self, *args, **kwargs):
timestamp = kwargs.pop('timestamp', None)
if timestamp is None:
timestamp = datetime.datetime.now(utc)
super(TimestampingStreamResult, self).status(
*args, timestamp=timestamp, **kwargs)
class _StringException(Exception):
"""An exception made from an arbitrary string."""
if not str_is_unicode:
def __init__(self, string):
if type(string) is not unicode:
raise TypeError(
"_StringException expects unicode, got %r" % (string,))
Exception.__init__(self, string)
def __str__(self):
return self.args[0].encode("utf-8")
def __unicode__(self):
return self.args[0]
# For 3.0 and above the default __str__ is fine, so we don't define one.
def __hash__(self):
return id(self)
def __eq__(self, other):
try:
return self.args == other.args
except AttributeError:
return False
def _format_text_attachment(name, text):
if '\n' in text:
return "%s: {{{\n%s\n}}}\n" % (name, text)
return "%s: {{{%s}}}" % (name, text)
def _details_to_str(details, special=None):
"""Convert a details dict to a string.
:param details: A dictionary mapping short names to ``Content`` objects.
:param special: If specified, an attachment that should have special
attention drawn to it. The primary attachment. Normally it's the
traceback that caused the test to fail.
:return: A formatted string that can be included in text test results.
"""
empty_attachments = []
binary_attachments = []
text_attachments = []
special_content = None
# sorted is for testing, may want to remove that and use a dict
# subclass with defined order for items instead.
for key, content in sorted(details.items()):
if content.content_type.type != 'text':
binary_attachments.append((key, content.content_type))
continue
text = content.as_text().strip()
if not text:
empty_attachments.append(key)
continue
# We want the 'special' attachment to be at the bottom.
if key == special:
special_content = '%s\n' % (text,)
continue
text_attachments.append(_format_text_attachment(key, text))
if text_attachments and not text_attachments[-1].endswith('\n'):
text_attachments.append('')
if special_content:
text_attachments.append(special_content)
lines = []
if binary_attachments:
lines.append('Binary content:\n')
for name, content_type in binary_attachments:
lines.append(' %s (%s)\n' % (name, content_type))
if empty_attachments:
lines.append('Empty attachments:\n')
for name in empty_attachments:
lines.append(' %s\n' % (name,))
if (binary_attachments or empty_attachments) and text_attachments:
lines.append('\n')
lines.append('\n'.join(text_attachments))
return _u('').join(lines)
|