ChapterPDF Available

Identifying Candidate Tasks for Robotic Process Automation in Textual Process Descriptions

Authors:

Figures

Content may be subject to copyright.
Identifying Candidate Tasks for Robotic Process
Automation in Textual Process Descriptions
Henrik Leopold, Han van der Aa, and Hajo A. Reijers
Vrije Universiteit Amsterdam
De Boelelaan 1081, 1081 HV Amsterdam, The Netherlands
h.leopold|j.h.vander.aa|h.a.reijers@vu.nl
Abstract. The continuous digitization requires organizations to im-
prove the automation of their business processes. Among others, this
has lead to an increased interest in Robotic Process Automation (RPA).
RPA solutions emerge in the form of software that automatically executes
repetitive and routine tasks. While the benefits of RPA on cost savings
and other relevant performance indicators have been demonstrated in
different contexts, one of the key challenges for RPA endeavors is to ef-
fectively identify processes and tasks that are suitable for automation.
Textual process descriptions, such as work instructions, provide rich and
important insights about this matter. However, organizations often main-
tain hundreds or even thousands of them, which makes a manual analysis
unfeasible for larger organizations. Recognizing the large manual effort
required to determine the current degree of automation in an organiza-
tion’s business processes, we use this paper to propose an approach that
is able to automatically do so. More specifically, we leverage supervised
machine learning to automatically identify whether a task described in
a textual process description is manual, an interaction of a human with
an information system or automated. An evaluation with a set of 424
activities from a total of 47 textual process descriptions demonstrates
that our approach produces satisfactory results.
1 Introduction
Many organizations currently face the challenge of keeping up with the increas-
ing digitization. Among others, it requires them to adapt existing business mod-
els and to respectively improve the automation of their business processes [29].
While the former is a rather strategic task, the latter calls for specific opera-
tional solutions. One of the most recent developments to increase the level of
automation is referred to as Robotic Process Automation (RPA). In essence,
RPA emerges in the form of software-based solutions that automatically execute
repetitive and routine tasks [5]. In this way, knowledge workers can dedicate
their time and effort to more complex and value adding tasks.
While the benefits of RPA have been demonstrated in different contexts
[8,21], one of the key challenges is to effectively identify processes and tasks that
are suitable for automation [6]. So far, research has focused on the establishment
2
of criteria [12, 38] and step-by-step guidelines [10] as means to support organi-
zations in addressing this challenge. However, what all these methods have in
common is that they require a manual analysis of the current degree of automa-
tion, i.e., they depend on the manual identification of tasks and (sub-)processes
that are automated or supported by an information system. This identifica-
tion task requires a thorough analysis of process-related documentations such as
process models and textual process documentations. While especially the latter
often provides rich and detailed insights, organizations typically maintain hun-
dreds or even thousands of them [3]. As a result, these methods do not scale for
organizations with hundreds of processes
Recognizing the large manual effort required to determine the current de-
gree of automation in an organization’s business processes, we use this paper
to propose an approach that is able to automatically do so. More specifically,
we combine supervised machine learning and natural language processing tech-
niques to automatically identify whether a task described in a textual process
description is a 1) manual task, 2) user task (interaction of a human with an
information system) or 3) automated task. An evaluation with a set of 424 activ-
ities from a total 47 textual process descriptions demonstrates that our approach
produces satisfactory results. Therefore, our approach can be employed to reduce
the effort required to determine the degree of automation in an organization’s
processes, as a first step in RPA endeavors.
The rest of the paper is organized as follows. Section 2 illustrates the problem
we address using a running example. Section 3 introduces our approach for auto-
matically determining the degree of automation of textual process descriptions
on a conceptual level. Section 4 presents the results of our evaluation. Section 5
discusses related work before Section 6 concludes the paper.
2 Problem Statement
In this section, we illustrate the problem of automatically identifying the degree
of automation of tasks described in a textual process description. Building on the
three categories of task automation introduced in [10], our goal is to classify each
task from a given textual process description as either 1) manual, 2) user task
(interaction of a human with an information system) or 3) automated. Figure 1
shows an exemplary textual process description, the associated relevant process
tasks, and their degree of automation.
Figure 1 shows that this textual process description contains two manual
tasks, two user tasks, and two automated tasks. The manual tasks include the
decision of the supervisor about the vacation request (task 3) and the completion
of the management procedures by the HR representative (task 6). The user tasks
are the two tasks in the process that are executed using the help of an information
system. That is, the submission and the reception of the vacation request (tasks
1 and 2). The automated tasks are tasks executed by the ERP system. This
includes returning the application to the employee (task 4) as well as generating
the notification to the HR representative (task 5). Analyzing this scenario in
3
The$ vaca(ons$ request$ process$ starts$ when$ an$ employee$ submits$ a$
vaca(on$request$ via$the$ERP$ system.$The$ request$ is$then$ received$by$the$
immediate$ supervisor$ of$ the$ employee$ reques(ng$ the$ vaca(on.$ The$
supervisor$ decides$ about$ the$ request.$ If$ the$ request$ is$ rejected,$ the$
applica(on$ is$ returned$ to$ the$ employee.$ If$ the$ request$ is$ approved,$ a$
no(fica(on$ is$ generated$ to$ the$ HR$ representa(ve,$ who$ then$ completes$
the$respec(ve$management$procedures.$
Task%ID:%1%
Verb:% %submit$
Object:%%vaca(on$request$
Role:% %employee$
Type: %user$task$
Task%ID:%2$
Verb:$ $receive$
Object:$$request$
Role:$ $supervisor$$
Type: $user$task$
Task%ID:%3$
Verb:% %decide$
Object:%%request$
Role:% %supervisor$
Type: %manual$
Task%ID:%4$
Verb:% %return$
Object:%%applica(on$
Role:% %ERP$system$$
Type: %automa(c$
Task%ID:%%5$
Verb:% %generate$
Object:%%no(fica(on$
Role:% %ERP$system%%
Type: %automa(c$
Task%ID:%6$
Verb:% %complete$
Object:%%mgmt.$procedures$
Role:% %HR$representa(ve$
Type: %manual$
Fig. 1. Process description with highlighted activities and their degree of automation
more detail, reveals that the automatic classification of these tasks is associated
with two main challenges:
1. Identification of tasks: Before a task can be classified, an automated ap-
proach must be able to detect the tasks described in a text. Note, for exam-
ple, that the verb “starts”, the verb “rejected” as well as the verb “approved
do not relate to tasks. The first is not relevant to the classification task at
hand because it represents a piece of meta information about the process.
The latter two tasks are not relevant because they rather relate to conditions
than to tasks, i.e., “if the request is rejected” describes a state, rather than
an activity being performed. Besides identifying relevant verbs, the identi-
fication of tasks also requires to properly infer the object to which a verb
refers and the resource that executes the task, i.e. the role.
2. Consideration of context: To reliably predict whether a certain activity is a
manual, user, or automated task, an automated approach must be able to
take a number of contextual factors into account. Consider, for instance, the
receipt of the vacation request (task 2). While in this process description the
request is submitted to an information system, this might not be the case
in other processes (a request could be also received orally or in writing).
The fact that an information system is mentioned in the first sentence, must
respectively be considered when classifying a task described later in the
process.
In prior work, only the former challenge has been addressed. The technique
for generating process models from natural language texts proposed by Friedrich
4
et al. [11] can reliably recognize and extract tasks from textual process descrip-
tions. To the best of our knowledge, there does not exist any technique that
addresses the second challenge. In this paper, we do so by operationalizing the
problem of automatically identifying the degree of task automation as multi-
class classification problem. In the next section, we elaborate on the details of
our proposed solution.
3 Conceptual Approach
In this section, we present our approach for automatically identifying the degree
of automation of tasks described in a textual process description. Section 3.1
first gives an overview of the approach. Section 3.2 introduces the dataset we
use in this paper, before sections 3.3 trough 3.5 elaborate on the details of our
approach.
3.1 Overview
The overall architecture of our three-step approach is visualized in Figure 2.
The approach takes as input a textual process descriptions and returns a list of
process tasks that are classified according to their degree of automation.
Classification
Textual Process
Description
Feature
Computation
Extracted
Features
Manually Labeled
Training Data
A
M
M
M
U
M
A
U
Automatically
Classified Activities
Linguistic
Preprocessing
Annotated
Textual Process
Description
Fig. 2. Overview of the proposed approach
The first step is to parse the text and to identify the relevant linguistic
entities and relations that denote tasks in a process description. For instance,
we determine which words represent verbs and to which objects they relate. The
result of this preprocessing step is a textual process description annotated with
the linguistic information related to the process’ tasks. The second step is the
computation of the features we use for prediction. In particular, we compute
features related to the verbs and objects that characterize tasks in a process, the
resources that execute tasks, and a feature characterizing terms from IT domains.
The output of this step is a feature table that contains the extracted tasks
and their corresponding features. In the third step, we perform a classification
5
based on the computed features. In the context of this paper, we use an SVM,
which is a supervised machine learning algorithm that automatically classifies
the input based on a set of manually labeled training instances. The output of
the classification is a list of tasks, each automatically classified as manual, user,
or automated.
In the following sections, we elaborate on each step in more detail. Because of
the supervised nature of our classification approach, we begin with introducing
our dataset.
3.2 Dataset
For this paper we use a subset of a collection of textual process descriptions
introduced in [11]. The collection contains 47 process descriptions from 10 dif-
ferent industrial and scholarly sources. We removed one of these sources (i.e.
14 process descriptions) because the textual descriptions from this source were
obtained using Google Translate and their language quality was insufficient for
our purposes. To obtain the required classifications for the 424 tasks described in
this dataset, two researchers independently classified each task as manual, user,
or automated. Conflicts were resolved by involving a third researcher. Table 1
gives an overview of the characteristics of the resulting dataset.
Table 1. Characteristics of dataset
ID Source Type D S W/S MT UT AT
1 HU Berlin Academic 4 10.0 18.1 52 4 1
2 TU Berlin Academic 2 34.0 21.2 42 38 11
3 QUT Academic 8 6.1 18.3 51 20 1
4 TU Eindhoven Academic 1 40.0 18.5 36 8 0
5 Vendor Tutorials Industry 4 9.0 18.2 9 23 2
6 inubit AG Industry 4 11.5 18.4 9 23 3
7 BPM Practitioners Industry 1 7 9.7 7 1 0
8 BPMN Practice Handbook Textbook 3 4.7 17.0 14 6 1
9 BPMN Guide Textbook 6 7.0 20.8 30 30 2
Total 33 9.7 16.8 250 153 21
Legend: D = Number of process descriptions per source, S = Average number
of sentences, W/S = Average number of words per sentence, MT = Total
number of manual tasks per source, UT = Total number of user tasks per
source, AT = Total number of automated tasks per source
The data from Table 1 illustrates that the process descriptions from our
dataset differ with respect to many dimensions. Most notably, they differ in size.
The average number of sentences ranges from 4.7 to 34.0. The longest process
description contains a total of 40 sentences. The descriptions also differ in the av-
erage length of the sentences. While the descriptions from the BPM Practitioners
6
source contain rather short sentences (9.7 words), the process descriptions from
the TU Berlin source contain relatively long sentences (21.2 words). The process
descriptions also differ with respect to the degree of automation. Some sources
contain process descriptions mostly covering manual tasks (e.g. the HU Berlin
source), others contain a quite considerable number of automated tasks (e.g. the
TU Berlin source). Lastly, the process descriptions differ in terms of how explic-
itly and unambiguously they describe the process behavior. Among others, this
results from the variety of authors that created the textual descriptions.
3.3 Linguistic Preprocessing
The goal of the linguistic preprocessing step is to automatically extract verbs,
object, and roles related to tasks described in the input text. To accomplish
this, we build on a technique that was originally developed for the extraction of
process models from natural language text [11]. This technique, which is regarded
as state-of-the-art [32], combines linguistic tools such as the Stanford Parser
[19] and VerbNet [34] to, among others, identify verbs, objects, and roles. The
advantage of this technique is its high accuracy and its ability to resolve so-called
anaphoric references such as “it” and “they ”. To illustrate the working principle
of the technique, consider the first sentence from the running example in Figure
1:
The vacations request process starts when an employee submits a vaca-
tion request via the ERP system.”
The first step is the application of the Stanford Parser, which automatically
detects the part of speech of each word as well as the grammatical relations
between them. The result of the part-of-speech tagging looks as follows.
The/DT vacations/NNS request/NN process/NN starts/VBZ when/WRB
an/DT employee/NN submits/VBZ a/DT vacation/NN request/NN vi-
a/IN the/DT ERP/NNP system/NN ./.
We can see that the Stanford Parser correctly identifies two verbs “starts
and “submits” (indicated by the tag “VBZ ”). The dependency analysis of the
Stanford Parser further reveals to which subjects and objects these verbs relate:
nsubj(starts-5, process-4)
nsubj(submits-9, employee-8)
dobj(submits-9, request-12)
compound(request-12, vacation-11)
The verb “starts” relates to the subject “process” and the verb “submits
relates the subject “employee” as well as the object “request”. The Stanford
Parser also recognizes that “vacation request” is a compound noun (i.e., a noun
that consists of several words). Based on the part-of-speech tagging output and
the dependency relations, the technique from [11] automatically extracts task
7
records consisting of a verb, an object, and the executing role. It also recognizes
that the verb “start” in this context represents meta information and not a
relevant task. It is respectively not included as a task record. The final set of
task records then represents the input to the next step of our approach.
3.4 Feature Computation
The selection and computation of suitable features is the key task when building
a machine learning-based solution [9]. Therefore, we manually analyzed which
characteristics in our dataset affect the degree of automation of a task. As a
result, we selected and implemented four features:
Verb feature (categorical)
Object feature (categorical)
Resource type (human/non-human)
IT domain (yes/no)
In the following paragraphs we elaborate on the definition and rationale of
each feature as well as its computation.
Verb feature The verb feature is a categorical feature and relates to the verb
used in the context of a task. The main idea behind this feature is that certain
verbs are more likely to be associated with automated tasks than others. As
an example, consider the verbs “generate” or “transmit”, which likely relate
to automated tasks. The verbs “analyze” and “decide”, by contrast, are more
likely to relate to manual tasks. The advantage of introducing a verb feature over
using predefined verb classed (such as the Levin verb classes [28]) is that a verb
feature does not tie a verb to a specific automation class. The verb “generate”,
for instance, might as well be used in the context of “generate ideas” and, thus,
refer to a manual task. Such a context-related use can be taken into account
when the verb is considered as part of a set of features.
The computation of this feature is straightforward since it is explicitly in-
cluded in the task record from the linguistic preprocessing step.
Object feature The object feature is a categorical feature and captures the
object that the verb of the task relates to. The rationale behind this feature is,
similar to the verb feature, that certain objects are more likely to be associated
with automated tasks than others. As an example, consider the two verb-object
combinations “send letter” and “send e-mail”. Although both contain the verb
send”, the object reveals that the former relates to a manual and the latter
relates to a user task (sending an e-email certainly requires the interaction with
a computer). While the number of objects we may encounter in textual process
descriptions is much higher than the number of verbs, including the object as a
feature might still help to differentiate different degrees of task automation.
Similar to the verb feature, the computation of this feature is straightforward
since it is part of the task record from the linguistic preprocessing step.
8
Resource type feature The resource type feature is a binary feature that char-
acterizes the resource executing a task as either “human” or “non-human ”. The
reason for encoding the resource as a binary feature instead of a classical categor-
ical feature is the high number of resources that can execute a task. Depending
on the domain of the considered process, resources may, among others, relate to
specific roles (e.g. “manager ” or “accountant”), departments (e.g., “HR depart-
ment” or “accounting department”), and also systems (“ERP system” or “infor-
mation system”). Despite this variety, the key characteristic revealing whether
a task is likely to be automated is the type of the resource, that is, whether the
resource is human or not. Apparently, a human resource can only relate to a
manual or user task, while a non-human resource can also execute automated
task (especially when the non-human resource represents an IT system).
Unlike the computation of the verb and the object feature, the computation
of the resource type feature is not trivial. The task record from the linguis-
tic preprocessing step only contains the actual resource and no indication of
the resource type. To determine the resource type, we use the lexical database
WordNet [30]. WordNet groups English words into sets of synonyms, so-called
synsets. For each of the 117,000 synsets WordNet contains, it provides short
definitions, examples, and a number of semantic relations to other synsets. To
compute this feature, we leverage the hypernym relationship from WordNet. In
general, a hypernym is a more generic term for a given word. For instance, the
word “vehicle” is the hypernym of “car” and the word “bird” is the hypernym
of “eagle”. Based on this notion of hypernymy and the hierarchical organization
of WordNet, we are able to infer for a given resource whether its hypernym is
physical entity”, “abstract entity” or a “person”. Based on this hypernym in-
formation we then can automatically categorize whether a resource is human or
non-human.
IT domain feature The IT domain feature is a binary feature that reveals
whether a task relates to the IT domain or not. The rationale behind this feature
is that a task that relates to the IT domain is likely to be a user task or even an
automated task. As example, consider the text fragment “the customer submits
a complaint via the complaint management system”. This fragment contains
the human actor “customer”, the verb “submit” and the object “complaint”.
Neither of these elements clearly indicates a degree of automation. However, the
fragment also mentions a “complaint management system ”. The goal of the IT
domain feature is to take such IT-related context into account.
To compute this feature, we leverage the glossary of computer terms devel-
oped by the University of Utah1. Besides a comprehensive coverage of technical
terms, this list also contains verbs and adjectives that are used in an IT context.
If a considered sentence, contains one or more terms from this list, the IT domain
feature receives the value “yes” for any task that is part of this sentence.
1http://www.math.utah.edu/ wisnia/glossary.html
9
3.5 Classification
In the final step of our approach, the actual classification of tasks from unseen
process descriptions takes place. As described in the previous section, there is
not a single feature that independently reveals the degree of automation of a
given task. It rather depends on the specific context of the task in the process.
To be able to still classify unseen tasks, we employ a Support Vector Machine [7],
a supervised machine learning algorithm. The advantages of SVMs are, among
others, that they can deal well with relatively small datasets, they have a low
risk of overfitting, and they scale well. For these reasons SVMs have also been
frequently applied in the context of other text classification tasks [17,36].
The core strategy of an SVM is to find a so-called hyperplane that best divides
a dataset. While in a two-dimensional space a simple line would be sufficient to
do so, an SVM maps the data to higher and higher dimensions until a hyperplane
can be formed that clearly segregates the data. Since an SVM is a supervised
machine learning algorithm, it needs to be trained on a manually labeled dataset.
In the next section, we describe how we implemented the approach outlined
in this section and demonstrate the effectiveness of the approach through a
quantitative evaluation.
4 Evaluation
This section reports on our evaluation experiments. We first elaborate on the
evaluation setup and the implementation of our approach. Then, we provide a
detailed discussion of the results.
4.1 Setup
The goal of the evaluation experiments is to demonstrate that the approach
introduced in this paper can reliably determine the degree of automation of
previously unseen textual process descriptions. To this end, we implemented our
approach as a Java prototype. Besides the code from [11], which we use to extract
tasks from a textual process descriptions, we build on the machine learning
library Weka [16] to implement the SVM, and JWNL [37] for incorporating and
accessing the lexical database WordNet.
To evaluate the performance of our approach, we conducted a repeated 10-
fold cross validation using our dataset. The idea behind this validation approach
is to randomly split the data set into 10 mutually exclusive subsets (so-called
folds) of about equal size. The SVM is then trained on 9 of the 10 folds and tested
on the remaining (unseen) fold. This process is repeated 10 times such that, in
the end, all data has been used for both training and testing. The advantage of
this evaluation method is that it does not require to partition the data set into
training and test data. We ran four different configurations of our approach:
1. Training on action feature only (A)
2. Training on action and object feature (A+O)
10
3. Training on action, object, and resource type feature (A+O+RT)
4. Training on all feature (Full)
To quantify the performance of each configuration , we use the standard met-
rics precision, recall, and F1-measure. In our context, precision for a particular
class is given by the number of tasks that were correctly assigned to this class
divided by the total number of tasks that were assigned to this class. Recall is
given by the number of tasks that were correctly assigned to this class divided by
the total number of tasks belong to that class. The F1-measure is the harmonic
mean of the two. Note that precision, recall, and F1-measure are computed for
each class individually. To also provide aggregate results, we conduct micro av-
eraging. That is, we use the number of tasks belonging to a particular class to
weight the respective precision and recall values. A macro perspective (i.e. ap-
plying no weights) would provide a distorted picture because the three classes
vary in size.
4.2 Results
The results of the 10-fold cross validation are presented in Table 2. Besides
precision, recall, F1-measure for each class and configuration, it also shows the
number of correctly and incorrectly classified instances.
A A+O A+O+RT Full
Correct 320 320 340 342
Incorrect 104 104 84 82
Manual
Precision 0.75 0.75 0.80 0.81
Recall 0.83 0.83 0.90 0.90
F1-Measure 0.79 0.79 0.85 0.85
User
Precision 0.75 0.75 0.80 0.80
Recall 0.67 0.67 0.70 0.70
F1-Measure 0.71 0.71 0.75 0.75
Automated
Precision 0.82 0.82 1.00 0.92
Recall 0.43 0.43 0.43 0.52
F1-Measure 0.56 0.56 0.60 0.66
Total (mic.)
Precision 0.75 0.75 0.80 0.81
Recall 0.75 0.75 0.80 0.80
F1-Measure 0.75 0.75 0.80 0.81
Table 2. Results from 10-fold cross validation
In general, the results from Table 2 reveal that our approach works well. Out
of the 424 task instances, our approach classified 342 correctly. This yields an
11
overall F1-measure of 0.81. Taking a look at the contribution of the individual
features shows that the action feature is of particular importance. Apparently
the discriminating power of the action feature is considerable, already resulting
in an overall F1-measure of 0.75. We can further see that adding the object
feature has no effect at all. However, the resource type feature results in a further
improvement of the overall F1-measure to 0.80. The IT domain feature has little
effect, but apparently leads to the correct classification of at least two additional
task instances.
Analyzing the results for individual classes in more detail shows that there
are quite some differences among the classes. Most notably, the F1-measure for
the automated class (0.66) is much lower than the F1-measure of the manual
(0.85) and the user class (0.75). This is, however, not particularly surprising
when taking the class sizes into account. The automated class only contains 21
instances, which clearly makes it a minority class. It is worth noting that the
rather low F1-measure mainly results from a low recall (0.52). The precision
reveals that automated task are correctly classified in 92% of the cases.
To further illustrate the results, Figure 3 shows the ROC curves and the cor-
responding AUC (area under the curve) values for the total configuration.2ROC
curves are graphical representations of the proportion of true positives versus the
proportion of false positives and often used to illustrate the capabilities of a bi-
nary classifier. The AUC value represents the probability that a classifier ranks
a randomly chosen positive instance higher than a randomly chosen negative
one. The AUC value varies between 0 and 1. An uninformative classifier yields
an AUC value of 0.5, a perfect classifier respectively yields an AUC value of 1.0.
The AUC values for our approach (ranging from 0.75 to 0.78 depending on the
class) indicate that our approach represents a classifier with a good performance.
To get insights into the limits of our approach, we conducted an error analy-
sis. More specifically, we investigated which task instances were classified incor-
rectly and why. In essence, we observed two main types of misclassifications: (1)
misclassifications due to a deviating use of feature attributes and (2) misclassi-
fications due to insufficient training data. The first category relates to instances
that were classified erroneously because the feature attributes are typically as-
sociated with another class. As an example, consider the manual task “attach
the new sct document”. For this task, our approach misses the fact that the “sct
document” is actually a physical document. It classifies it as a user task because
the verb “send” is often associated with user tasks in our dataset (e.g. consider
send e-mail”). Another example is the user task “check notes”, which our ap-
proach classified as a manual task. Here it did not recognize the context of an
information system and bases its decision on the verb “check” and the object
notes”, which are often associated with manual tasks. The second category of
misclassifications relates to cases where our approach erroneously classified a
2Note that the way Weka generates ROC curves results in only as many threshold val-
ues as there are distinct probability values assigned to the positive class. Therefore,
the ROC curves from Figure 3 are only based on three data points. This, however,
does not reduce their informative value.
12
!"
!#$"
!#%"
!#&"
!#'"
!#("
!#)"
!#*"
!#+"
!#,"
$"
!" !#%" !#'" !#)" !#+" $"
True%Posi*ve%Rate%
False%Posi*ve%Rate%
-./"/0123"451"/6788"9:7;0769"<=>/"?"!#*+@"
-./"/0123"451"/6788">8319"<=>/"?"!#*+@"
-./"/0123"451"/6788"9=0A5B7CD9"<=>/"?"!#*(@"
Fig. 3. ROC Curves for each class
task because it has not seen enough training data. For example, consider the
user task “transmit a response comment ”, which our approach classified as a
manual task. Here the problem is that our approach has not observed a suffi-
cient number of instances using the verb “transmit”, which clearly relates to the
use of an information system.
Despite these misclassifications, we can state that the presented approach
represents a promising solution for automatically determining the degree of task
automation.
5 Related Work
This paper relates to two major streams of research: (1) the application of Natu-
ral Language Process (NLP) technology in the context of business process anal-
ysis and (2) process automation.
A variety of authors have applied NLP technology in the context of business
process analysis. Their works can be subdivided into techniques that analyze
the natural language inside process models and techniques that analyze the
natural language outside of process models, typically captured in textual process
descriptions. Techniques analyzing the natural language inside process models
typically focus on activity and event labels. Among others, there exist techniques
for checking the correctness and consistency of process model labels [24, 27, 31],
techniques for identifying similar parts of process models [20,26], and techniques
for inferring information from process models such as service candidates [13,25].
Other approaches focus on the analysis of process-related text documents, such
as approaches for the automated elicitation of process models from texts, cf. [1,
11, 15] and the comparison of natural language texts to process models [2, 33],
and process querying based on the analysis of textual process descriptions [23].
13
The focus on automation in the context of Business Process Management is
not a recent development. In particular research on workflow management and
automation reaches back over 20 years [4,14,35]. Research on RPA, by contrast, is
still relatively scarce. Lacity and Willcocks investigated how organizations apply
RPA in practice [21, 22, 38]. They found that most applications of RPA have
been done for automating tasks of service business processes, such as validating
the sale of insurance premiums, generating utility bills, and keeping employee
records up-to date. Their study also revealed the overall potential of RPA ranging
from a significant increase in turnaround times and greater workforce flexibility
to cost savings of up to 30%. Other authors also studied the risks associated
with BPA. For instance, Kirchmer [18] argues that RPA has the potential to
make mistakes faster and with higher certainty because there is often no human
check before executing an action. Davenport and Kirby also tried to answer the
question of what machines are currently capable of. They argue that there are
four levels of intelligence that machines can potentially master: (1) support for
humans, (2) repetitive task automation, (3) context awareness and learning, and
(4) self-awareness. Currently, they conclude, machines are capable of mastering
level 1 and 2. Level 3 is only covered to a limited extend, level 4 not at all.
They, however, stress that machines are advancing and that it is important to
understand how human capabilities fit into the picture. [8].
6 Conclusion
In this paper, we proposed a machine learning-based approach that automati-
cally identifies and classifies tasks from textual process descriptions as manual,
user, or automated. The goal of our technique is to reduce the effort that is
required to identify suitable candidates for robotic process automation. An eval-
uation with 424 activities from a total of 47 textual process descriptions showed
that our approach achieves an F-measure of 0.81 and, therefore, produces satis-
factory results.
Despite these positive results, it is important to consider our results in the
light of some limitations. First, it should be noted that the dataset we used in this
paper is not representative. Textual process descriptions in practice may deviate
from the ones in our dataset in different ways. However, we tried to maximize the
external validity of our evaluation by choosing a dataset that combines different
sources. What is more, our approach could be easily retrained on other datasets
to further increase its performance. Second, our approach cannot guarantee that
suitable automation candidates are identified. It rather gives an overview of the
current degree of automation, which can then serve as input for further analyses.
In future work, we plan to improve the performance of our approach by
including additional features and testing other classifiers. What is more, we
intend to apply our approach in organizations in order to obtain feedback about
its usefulness in practice.
14
References
1. Van der Aa, H., Leopold, H., Reijers, H.A.: Dealing with behavioral ambiguity
in textual process descriptions. In: International Conference on Business Process
Management. pp. 271–288. Springer (2016)
2. Van der Aa, H., Leopold, H., Reijers, H.A.: Comparing textual descriptions to
process models: The automatic detection of inconsistencies. Information Systems
64, 447–460 (2017)
3. Van der Aa, H., Leopold, H., van de Weerd, I., Reijers, H.A.: Causes and conse-
quences of fragmented process information: Insights from a case study. In: Pro-
ceedings of the the annual Americas’ Conference on Information Systems (2017)
4. van der Aalst, W.M., Barros, A.P., ter Hofstede, A.H., Kiepuszewski, B.: Advanced
workflow patterns. In: International Conference on Cooperative Information Sys-
tems. pp. 18–29. Springer (2000)
5. Aguirre, S., Rodriguez, A.: Automation of a business process using robotic pro-
cess automation (rpa): A case study. In: Figueroa-Garc´ıa, J.C., L´opez-Santana,
E.R., Villa-Ram´ırez, J.L., Ferro-Escobar, R. (eds.) Applied Computer Sciences in
Engineering. pp. 65–71. Springer International Publishing, Cham (2017)
6. Aguirre, S., Rodriguez, A.: Automation of a business process using robotic process
automation (rpa): A case study. In: Workshop on Engineering Applications. pp.
65–71. Springer (2017)
7. Cortes, C., Vapnik, V.: Support-vector networks. Machine learning 20(3), 273–297
(1995)
8. Davenport, T.H., Kirby, J.: Just how smart are smart machines? MIT Sloan Man-
agement Review 57(3), 21 (2016)
9. Domingos, P.: A few useful things to know about machine learning. Communica-
tions of the ACM 55(10), 78–87 (2012)
10. Dumas, M., Rosa, M., Mendling, J., Reijers, H.: Fundamentals of Business Process
Management. Springer (2013)
11. Friedrich, F., Mendling, J., Puhlmann, F.: Process Model Generation from Natural
Language Text. In: Proceedings of the 23rd international conference on Advanced
Information Systems Engineering. LNCS, vol. 6741, pp. 482–496. Springer (2011)
12. Fung, H.P.: Criteria, use cases and effects of information technology process au-
tomation (itpa). Browser Download This Paper (2014)
13. Gacitua-Decar, V., Pahl, C.: Automatic business process pattern matching for
enterprise services design. Services Part II, IEEE Congress on 0, 111–118 (2009)
14. Georgakopoulos, D., Hornick, M., Sheth, A.: An overview of workflow management:
From process modeling to workflow automation infrastructure. Distributed and
parallel Databases 3(2), 119–153 (1995)
15. Ghose, A.K., Koliadis, G., Chueng, A.: Process Discovery from Model and Text
Artefacts. In: Proceedings of the IEEE Congress on Services. pp. 167–174. IEEE
Computer Society (2007)
16. Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., Witten, I.H.: The
weka data mining software: an update. ACM SIGKDD explorations newsletter
11(1), 10–18 (2009)
17. Joachims, T.: Text categorization with support vector machines: Learning with
many relevant features. In: European conference on machine learning. pp. 137–
142. Springer (1998)
18. Kirchmer, M.: Robotic process automation–pragmatic solution or dangerous illu-
sion? BTOES Insights, June (2017)
15
19. Klein, D., Manning, C.D.: Accurate Unlexicalized Parsing. 41st Meeting of the
Association for Computational Linguistics pp. 423–430 (2003)
20. Klinkm¨uller, C., Weber, I., Mendling, J., Leopold, H., Ludwig, A.: Increasing recall
of process model matching by improved activity label matching. In: BPM, pp. 211–
218. Springer (2013)
21. Lacity, M., Willcocks, L.P., Craig, A.: Robotic process automation at telefonica o2
(2015)
22. Lacity, M.C., Willcocks, L.P.: A new approach to automating services. MIT Sloan
Management Review 58(1), 41 (2016)
23. Leopold, H., van der Aa, H., Pittke, F., Raffel, M., Mendling, J., Reijers, H.A.:
Searching textual and model-based process descriptions based on a unified data
format. Software & Systems Modeling pp. 1–16 (2017)
24. Leopold, H., Eid-Sabbagh, R.H., Mendling, J., Azevedo, L.G., Bai˜ao, F.A.: De-
tection of naming convention violations in process models for different languages.
Decision Support Systems 56(0), 310–325 (12 2013)
25. Leopold, H., Mendling, J.: Automatic derivation of service candidates from business
process model repositories. In: Business Information Systems. pp. 84–95 (2012)
26. Leopold, H., Niepert, M., Weidlich, M., Mendling, J., Dijkman, R.M., Stucken-
schmidt, H.: Probabilistic optimization of semantic process model matching. In:
BPM. pp. 319–334 (2012)
27. Leopold, H., Smirnov, S., Mendling, J.: On the refactoring of activity labels in
business process models. Information Systems 37(5), 443–459 (2012)
28. Levin, B.: English Verb Classes and Alternations: A Preliminary Investigation.
University of Chicago Press (1993)
29. Leyh, C., Bley, K., Seek, S.: Elicitation of processes in business process manage-
ment in the era of digitization – the same techniques as decades ago? In: Piazolo,
F., Geist, V., Brehm, L., Schmidt, R. (eds.) Innovations in Enterprise Information
Systems Management and Engineering. pp. 42–56. Springer International Publish-
ing, Cham (2017)
30. Miller, G., Fellbaum, C.: WordNet: An Electronic Lexical Database. MIT Press,
Cambridge, MA (1998)
31. Pittke, F., Leopold, H., Mendling, J.: Automatic detection and resolution of lexical
ambiguity in process models (2015)
32. Riefer, M., Ternis, S.F., Thaler, T.: Mining process models from natural lan-
guage text: A state-of-the-art analysis. In: Multikonferenz Wirtschaftsinformatik
(MKWI-16), March 9-11, Illmenau, Germany. Universit¨at Illmenau (2016)
33. S`anchez-Ferreres, J., Carmona, J., Padr´o, L.: Aligning textual and graphical de-
scriptions of processes through ILP techniques. In: International Conference on
Advanced Information Systems Engineering (In press). Springer (2017)
34. Schuler, K.K.: Verbnet: a broad-coverage, comprehensive verb lexicon. Ph.D. thesis,
Philadelphia, PA, USA (2005)
35. Stohr, E.A., Zhao, J.L.: Workflow automation: Overview and research issues. In-
formation Systems Frontiers 3(3), 281–296 (2001)
36. Tong, S., Koller, D.: Support vector machine active learning with applications to
text classification. Journal of machine learning research 2(Nov), 45–66 (2001)
37. Walenz, B., Didion, J.: Jwnl: Java wordnet library (2011)
38. Willcocks, L., Lacity, M.C.: Service automation: Robots and the future of work.
Steve Brookes Publishing (2016)
... Another type of environmental change that was an important driver for adaptations of business processes in the past is technological change. For instance, methods have been introduced to automatically identify tasks applicable to Robotic Process Automation (Leopold et al. 2018) or outsourcing to the Cloud (Rekik et al. 2015). Due to the heterogeneous nature of environmental changes and the specific requirements for each automatic identification, the concrete steps as well as the implemented logic in these methods differ significantly (Nake et al. 2024). ...
... • Gather Data for Classification: Afterwards, data for classification must be gathered. This can be textual business process descriptions that often already exist in companies (Leopold et al. 2018) or other text fragments related to a business process; the text in business process models, for example. ...
Conference Paper
Full-text available
Organizations constantly face changes arising from the environment and from within. The degree of environmental change has increased due to technological innovation and other factors, like the introduction of new regulations. These changes directly influence the business processes of organizations and often make transformations necessary. To carry out such transformations, the first step is to identify the tasks involved. Several automatic methods have been introduced in the literature to identify tasks affected by specific changes. Still, general methods that can be applied to different types of change are lacking. The introduction of new technologies, like large language models, makes such general methods feasible. Consequently, in this research, we first present the design requirements and principles for a procedural model for automatic context-specific task identification based on insights from the literature and propose such a model. Finally, we conduct an initial evaluation and derive the next steps for this research.
... This section reviews relevant literature on the automated generation of BPMN process models from textual descriptions using NLP techniques. Approaches for extracting BPMN models from textual descriptions using NLP can be grouped as follows: Rule-based approaches [1], [2], [3], [4], [5], [6], [7], [8] Machine learning and deep learning-based approaches [9], [10], [11], [13] Machine translation approaches [15] Constrained natural language approaches [16] ...
... In machine learning and deep learning, relevant studies can be categorized into task classification and process model extraction approaches. Leopold et al. [9] propose a machine learning-based method for task classification using linguistic features and support vector machines. In contrast, Qian et al. [10] developed a deep learning-based hierarchical neural network for multi-grained text classification, capturing procedural knowledge. ...
... With regard to the second stream, contemporary research has put RPA in the context of traditional business process management, mainly focusing on achieving instrumental goals of improvements regarding time, cost, quality, and flexibility (François et al. 2022). In this context, a lot of effort has been invested into identifying characteristics of RPA-suitable tasks (Syed et al. 2020) and identifying them in an automated way, e.g., from textual process descriptions (Leopold et al. 2018). To this end, prior research used Fig. 1 The current field of RPA research predominantly behavioral or design-oriented methods to create both descriptive and prescriptive insights (Plattfaut and Borghoff 2022). ...
Article
Full-text available
Robotic Process Automation is an established technology in organizations. In the last years, it has also received considerable attention in scholarly research with publications, special issues, and academic conferences dedicated to the topic. Given that Robotic Process Automation has now moved beyond the initial hype, we can ask what research should focus on in the future. To address this question, we conducted a panel discussion to discuss its current state and future development. This panel, which took place at the Robotic Process Automation forum at the Business Process Management Conference 2023, included experts from academia and industry, covering strategy consultants, implementers, and tool providers. In this report, we present insights from the panel discussions. We especially focus on three future research directions on Robotic Process Automation that emerged from the panel.
... Traditionally, ML models in the sub-discipline of NLP are designed to process natural language texts and automatically extract their meaning [38]. NLP can be used for example not only in the execution phase to categorize user inquiries based on user sentiment, but it can be used also in other phases such as process selection for identifying automation candidates [26]. ...
... Robotic Process Automation (RPA) aims to automate business processes or parts of them with software robots, reproducing human interactions with the graphical user interface [2]. Software robots make it possible to automate Backoffice/Administrative related jobs, which were first performed by human collaborators [3]. Recently, many RPA approaches have been implemented and the RPA software market grew by 60% in 2018 [4]. ...
Article
Full-text available
p>Robotic process automation (RPA) is a cutting-edge technology that provides software robots to repeat and mimic the repeatable tasks that a human user earlier performed. The use of software robots is encouraging because of their cost efficiency and easy implementation. Selecting and prioritizing a candidate process for automation is always challenging as all the business processes in an organization are not equally suitable for RPA implementation. Various studies have highlighted several criteria found in the literature for determining, prioritising, and selecting a business process for RPA. Nevertheless, there are no set standards for evaluating and analyzing a certain process or its tasks to determine whether they may be automated to use RPA. This paper aims to develop standard criteria and propose a consistent model to select and prioritize candidate process for RPA projects. To assess these criteria's applicability in the context of RPA, surveys among subject matter experts (SMEs) are used to validate them. Principal component analysis (PCA) and correlation are used to identify the top 20 criteria. Naïve Bayes algorithm is applied on the collected data for decision-making. The developed multi-criteria model exhibits strong precision and recall measures, with training and validation accuracy of 96% and 90%, respectively.</p
Chapter
Due to the 4th Industrial Revolution, Artificial Intelligence and other internet technologies such as RAISA (Robots, Artificial Intelligence, and Service Automation) have been adopted in various industries, including the hospitality and tourism industry. As the industry traditionally offers services through a manual labor force, more companies adopt RAISA through a commercial service in the form of chatbots, delivery robots, robot concierge, conveyor restaurants, self-service information/check-in/check-out kiosks, and many others. This book chapter explores the dynamic interplay between consumer behavior, customer orientation, emotional intelligence, and the integration of service robots into the customer service landscape. It delves into the transformative potential of service robots in various industries and the profound impact they have on customer satisfaction. In this chapter used holistic approach to examining the intersection of technology and human empathy in the context of customer experiences, offering insights into how service robots can enhance, rather than replace, human interactions within the customer service industry.The chapter also examines how understanding consumer behavior, cultivating customer orientation, and leveraging emotional intelligence can maximize the effectiveness of service robots in enhancing customer experiences. The integration of these components is crucial in achieving a harmonious synergy between technology and human interaction, ultimately shaping the future of customer service.
Article
Full-text available
Documenting business processes using process models is common practice in many organizations. However, not all process information is best captured in process models. Hence, many organizations complement these models with textual descriptions that specify additional details. The problem with this supplementary use of textual descriptions is that existing techniques for automatically searching process repositories are limited to process models. They are not capable of taking the information from textual descriptions into account and, therefore, provide incomplete search results. In this paper, we address this problem and propose a technique that is capable of searching textual as well as model-based process descriptions. It automatically extracts activity-related and behavioral information from both descriptions types and stores it in a unified data format. An evaluation with a large Austrian bank demonstrates that the additional consideration of textual descriptions allows us to identify more relevant processes from a repository.
Conference Paper
Full-text available
Robotic Process Automation (RPA) emerges as software based solution to automate rules-based business processes that involve routine tasks, structured data and deterministic outcomes. Recent studies report the benefits of the application of RPA in terms of productivity, costs, speed and error reduction. Most of these applications were carried out on back office business process where the customer is not directly involved, therefor a case study was conducted on a BPO provider to verify the benefits and results of applying RPA to a service business process with front and back office activities. The results show that productivity improvement is the main benefit of RPA, nevertheless time reduction was not achieved on this case.
Article
Full-text available
Robotic Process Automation (RPA) has become an important trend in process management and automation. The article examines briefly the value it delivers and associated risks. On this basis it recommends a company specific approach to maximize RPA value and minimize risk.
Conference Paper
Full-text available
Having access to the right information is vital to the effective and efficient execution of an organization's business processes. A major challenge in this regard is that information on a single process is often spread out over numerous models, documents, and systems. Despite the potential consequences of this situation, there is a lack of insights on how to mitigate its impact. Against this background, we conducted an explorative case study to analyze the causes and consequences of the fragmentation of process information. We found that the widespread fragmentation of information had a considerable impact on the investigated organization. In particular, fragmentation led to severe maintenance issues, reduced process execution efficiency, and had a negative effect on the quality of process results. Our findings provide useful insights for both practice and research on how to mitigate the negative aspects associated with the fragmentation of process information.
Article
Business Process Management (BPM) is the art and science of how work should be performed in an organization in order to ensure consistent outputs and to take advantage of improvement opportunities, e.g. reducing costs, execution times or error rates. Importantly, BPM is not about improving the way individual activities are performed, but rather about managing entire chains of events, activities and decisions that ultimately produce added value for an organization and its customers. This textbook encompasses the entire BPM lifecycle, from process identification to process monitoring, covering along the way process modelling, analysis, redesign and automation. Concepts, methods and tools from business management, computer science and industrial engineering are blended into one comprehensive and inter-disciplinary approach. The presentation is illustrated using the BPMN industry standard defined by the Object Management Group and widely endorsed by practitioners and vendors worldwide. In addition to explaining the relevant conceptual background, the book provides dozens of examples, more than 100 hands-on exercises – many with solutions – as well as numerous suggestions for further reading. The textbook is the result of many years of combined teaching experience of the authors, both at the undergraduate and graduate levels as well as in the context of professional training. Students and professionals from both business management and computer science will benefit from the step-by-step style of the textbook and its focus on fundamental concepts and proven methods. Lecturers will appreciate the class-tested format and the additional teaching material available on the accompanying website fundamentals-of-bpm.org.
Conference Paper
With the aim of having individuals from different backgrounds and expertise levels examine the operations in an organization, different representations of business processes are maintained. To have these different representations aligned is not only a desired feature, but also a real challenge due to the contrasting nature of each process representation. In this paper we present an efficient technique for aligning a textual description and a graphical model of a process. The technique is grounded on using natural language processing techniques to extract linguistic features of each representation, and encode the search as a mathematical optimization encoded using Integer Linear Programming (ILP) whose resolution ensures an optimal alignment between both descriptions. The technique has been implemented and the experiments witness the significance of the approach with respect to the state-of-the-art technique for the same task.
Conference Paper
For many decades, process models have built the basis for economically successful participation in the market. Companies are still faced with the task of identifying, defining and visualizing their processes, especially in today’s era of digitization. In this era and due to more and more complex inter-organizational processes across corporate boundaries, the question arises as whether techniques and approaches for elicitation of processes in the context of business process management have changed or if the established techniques are still appropriate. Here, digitization could have significant potential to automate the elicitation of processes. To address this issue we have conducted a systematic literature review and identified the theoretical requirements for the elicitation of processes. Then, based on an interview study with experienced consultants, we compared the theoretic results with the current applied techniques in today’s practice. Selected results are presented and discussed in this paper.
Article
Thesupport-vector network is a new learning machine for two-group classification problems. The machine conceptually implements the following idea: input vectors are non-linearly mapped to a very high-dimension feature space. In this feature space a linear decision surface is constructed. Special properties of the decision surface ensures high generalization ability of the learning machine. The idea behind the support-vector network was previously implemented for the restricted case where the training data can be separated without errors. We here extend this result to non-separable training data.High generalization ability of support-vector networks utilizing polynomial input transformations is demonstrated. We also compare the performance of the support-vector network to various classical learning algorithms that all took part in a benchmark study of Optical Character Recognition.
Article
While many white collar workers may feel threatened by service automation, companies that thoughtfully automate services are finding that the worries are overblown. By pairing humans and robots, companies can deliver better services for less, and jobs can become more interesting.