Abstract and Compare: A Framework for Defining Precision Measures for Automated Process Discovery
AuthorAugusto, A; Armas Cervantes, A; Conforti, R; Dumas, M; La Rosa, M; Reissner, D
Source TitleLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
University of Melbourne Author/sConforti, Raffaele; Augusto, Adriano; Armas Cervantes, Abel; La Rosa, Marcello; Reissner, Daniel; Augusto, Adriano
AffiliationComputing and Information Systems
Document TypeConference Paper
CitationsAugusto, A., Armas Cervantes, A., Conforti, R., Dumas, M., La Rosa, M. & Reissner, D. (2018). Abstract and Compare: A Framework for Defining Precision Measures for Automated Process Discovery. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 11080 LNCS, pp.158-175. SpringerLink. https://doi.org/10.1007/978-3-319-98648-7_10.
Access StatusOpen Access
ARC Grant codeARC/DP180102839
Automated process discovery techniques allow us to extract business process models from event logs. The quality of process models discovered by these techniques can be assessed with respect to various quality criteria related to simplicity and accuracy. One of these criteria, namely precision, captures the extent to which the behavior allowed by a discovered process model is observed in the log. While numerous measures of precision have been proposed in the literature, a recent study has shown that none of them fulfils a set of five axioms that capture intuitive properties behind the concept of precision. In addition, several existing precision measures suffer from scalability issues when applied to models discovered from real-life event logs. This paper presents a versatile framework for defining precision measures based on behavior abstractions. The key idea is that a precision measure can be defined by three ingredients: a function that abstracts a process model (e.g. as a transition system), a function that does the same for an event log, and a function that compares the behavior abstraction of the model with that of the log. We show empirically that different instances of this framework allow us to strike different tradeoffs between scalability and sensitivity. We also show that two instances of the framework based on lossless abstraction functions yield a precision measure that fulfils all the above-mentioned axioms.
- Click on "Export Reference in RIS Format" and choose "open with... Endnote".
- Click on "Export Reference in RIS Format". Login to Refworks, go to References => Import References