We currently don't have a record of tests executed, the platforms and environments they were run upon, and what code they covered. The only indirect evidence is available in open bugs, when people care to fill in these details. :-(
=Currently used tests and frameworks=
== Manual test plan ==
See [http://sourceforge.net/p/gramps/source/ci/master/tree/TestPlan.txt TestPlan.txt] in gramps toplevel. (I believe this is only done at a major release (like 4.0.0)).
* Unify running all the tests, in- or out- tree
** Try switching the feature tests and non-automated unit tests in test/ from our runner scripts to [http://docs.python.org/2/library/unittest.html#test-discovery python native unittest discovery mechanism].
** We have a stale [http://sourceforge.net/p/gramps/source/ci/master/tree/gramps/test/regrtest.py gramps/test/regrtest.py] runner, notable for logging init. - should we revive that, or maybe integrate into setup.py test runner?
* [https://wiki.python.org/moin/CodeCoverage| Coverage analysis]
* Continuous test status report, coverage, automatic deployment into win/mac/linux VMs (needs server capacity to be hosted online)(I can dream, can't I?)
* Automated regression tests for our GUI. The following links look interesting: