Picture of UK Houses of Parliament

Leading national thinking on politics, government & public policy through Open Access research

Strathprints makes available scholarly Open Access content by researchers in the School of Government & Public Policy, based within the Faculty of Humanities & Social Sciences.

Research here is 1st in Scotland for research intensity and spans a wide range of domains. The Department of Politics demonstrates expertise in understanding parties, elections and public opinion, with additional emphases on political economy, institutions and international relations. This international angle is reflected in the European Policies Research Centre (EPRC) which conducts comparative research on public policy. Meanwhile, the Centre for Energy Policy provides independent expertise on energy, working across multidisciplinary groups to shape policy for a low carbon economy.

Explore the Open Access research of the School of Government & Public Policy. Or explore all of Strathclyde's Open Access research...

Automatically classifying test results by semi-supervised learning

Almaghairbe, Rafig and Roper, Marc (2016) Automatically classifying test results by semi-supervised learning. In: 2016 IEEE 27th International Symposium on Software Reliability Engineering (ISSRE). IEEE, [Piscataway, NJ], pp. 116-126. ISBN 978-1-4673-9003-3

[img]
Preview
Text (Almaghairbe-Roper-ISSRE2016-Automatically-classifying-test-results-by-semi-supervised-learning)
Almaghairbe_Roper_ISSRE2016_Automatically_classifying_test_results_by_semi_supervised_learning.pdf
Accepted Author Manuscript

Download (189kB)| Preview

    Abstract

    A key component of software testing is deciding whether a test case has passed or failed: an expensive and error-prone manual activity. We present an approach to automatically classify passing and failing executions using semi-supervised learning on dynamic execution data (test inputs/outputs and execution traces). A small proportion of the test data is labelled as passing or failing and used in conjunction with the unlabelled data to build a classifier which labels the remaining outputs (classify them as passing or failing tests). A range of learning algorithms are investigated using several faulty versions of three systems along with varying types of data (inputs/outputs alone, or in combination with execution traces) and different labelling strategies (both failing and passing tests, and passing tests alone). The results show that in many cases labelling just a small proportion of the test cases – as low as 10% – is sufficient to build a classifier that is able to correctly categorise the large majority of the remaining test cases. This has important practical potential: when checking the test results from a system a developer need only examine a small proportion of these and use this information to train a learning algorithm to automatically classify the remainder.