Thu 7 Dec 2023 11:15 - 11:30 at Golden Gate C1 - Testing IV Chair(s): Jonathan Bell

The Cancer Registry of Norway (CRN) collects, curates, and manages data related to cancer patients in Norway, supported by an interactive, human-in-the-loop, socio-technical decision support software system. Automated software testing of this software system is inevitable; however, currently, it is limited in CRN’s practice. To this end, we present an industrial case study to evaluate an AI-based system-level testing tool, i.e., EvoMaster, in terms of its effectiveness in testing CRN’s software system. In particular, we focus on GURI, CRN’s medical rule engine, which is a key component at the CRN. We test GURI with EvoMaster’s black-box and white-box tools and study their test effectiveness regarding code coverage, errors found, and domain-specific rule coverage. The results show that all EvoMaster tools achieve a similar code coverage; i.e., around 19% line, 13% branch, and 20% method; and find a similar number of errors; i.e., 1 in GURI’s code. Concerning domain-specific coverage, EvoMaster’s black-box tool is the most effective in generating tests that lead to applied rules; i.e., 100% of the aggregation rules and between 12.86% and 25.81% of the validation rules; and to diverse rule execution results; i.e., 86.84% to 89.95% of the aggregation rules and 0.93% to 1.72% of the validation rules pass, and 1.70% to 3.12% of the aggregation rules and 1.58% to 3.74% of the validation rules fail. We further observe that the results are consistent across 10 versions of the rules. Based on these results, we recommend using EvoMaster’s black-box tool to test GURI since it provides good results and advances the current state of practice at the CRN. Nonetheless, EvoMaster needs to be extended to employ domain-specific optimization objectives to improve test effectiveness further. Finally, we conclude with lessons learned and potential research directions, which we believe are applicable in a general context.

Thu 7 Dec

Displayed time zone: Pacific Time (US & Canada) change

11:00 - 12:30
Testing IVResearch Papers / Industry Papers at Golden Gate C1
Chair(s): Jonathan Bell Northeastern University
11:00
15m
Talk
Code Coverage Criteria for Asynchronous Programs
Research Papers
Mohammad Ganji Simon Fraser University, Saba Alimadadi Simon Fraser University, Frank Tip Northeastern University
Media Attached
11:15
15m
Talk
Automated Test Generation for Medical Rules Web Services: A Case Study at the Cancer Registry of Norway
Industry Papers
Christoph Laaber Simula Research Laboratory, Tao Yue Beihang University, Shaukat Ali Simula Research Laboratory and Oslo Metropolitan University, Thomas Schwitalla Cancer Registry of Norway, Jan F. Nygård Cancer Registry of Norway
DOI Pre-print Media Attached
11:30
15m
Talk
Test Case Generation for Drivability Requirements of an Automotive Cruise Controller: An Experience with an Industrial Simulator
Industry Papers
Federico Formica McMaster University, Nicholas Petrunti McMaster University, Lucas Bruck McMaster University, Vera Pantelic McMaster University, Mark Lawford McMaster University, Claudio Menghi University of Bergamo; McMaster University
DOI Media Attached
11:45
15m
Talk
Prioritizing Natural Language Test Cases Based on Highly-Used Game Features
Industry Papers
Markos Viggiato University of Alberta, Dale Paas Prodigy Education, Cor-Paul Bezemer University of Alberta
DOI Media Attached
12:00
15m
Talk
EtherDiffer: Differential Testing on RPC Services of Ethereum Nodes
Research Papers
Shinhae Kim The Affiliated Institute of ETRI, Sungjae Hwang Sungkyunkwan University
Media Attached
12:15
15m
Talk
[Remote] API-Knowledge Aware Search-based Software Testing: Where, What and How
Research Papers
Xiaoxue Ren Zhejiang University, Xinyuan Ye Australian National University, Yun Lin Shanghai Jiao Tong University, Zhenchang Xing Data61, Shuqing Li The Chinese University of Hong Kong, Michael Lyu The Chinese University of Hong Kong
Media Attached