<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Business Process Improvement with Performance-Based Sequential Experiments (Extended Abstract)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Suhrid Satyal</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of New South Wales</institution>
          ,
          <country country="AU">Australia</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Various life-cycle approaches to Business Process Management (BPM) have a common assumption that a process is incrementally improved in the redesign phase. While this assumption is hardly questioned in BPM research, there is evidence from the eld of AB testing that improvement concepts often do not lead to actual improvements. In this thesis, we propose a methodology named AB-BPM and a set of supporting techniques that facilitate rapid validation of business processes by conducting sequential experiments. We evaluate our methodology and techniques with real-world and synthetic case studies.</p>
      </abstract>
      <kwd-group>
        <kwd>Suhrid Satyal</kwd>
        <kwd>Business process management</kwd>
        <kwd>AB Testing</kwd>
        <kwd>Shadow Testing</kwd>
        <kwd>Trace simulation</kwd>
        <kwd>Process performance indicators</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Business process improvement ideas often do not lead to actual
improvements.Works on business improvement ideas found that only a third of
the ideas observed had a positive impact [
        <xref ref-type="bibr" rid="ref3 ref4">3,4</xref>
        ] This is also illustrated
by an anecdote of a European Bank. The bank improved their loan
approval process by cutting its turnaround time down from one week to a
few hours as a means to boost their business. What happened though
was a steep decline in customer satisfaction: customers with a negative
notice would complain that their application might have been declined
unjusti ably; customers with a positive notice would inquire whether
their application had been checked with due diligence. These
observations emphasize the need to carefully test improvement hypotheses in
practice because the customers and the process participants might not
act as anticipated by the process analyst.
      </p>
      <p>
        Contemporary BPM research does not provide techniques and guidelines
on quickly testing and validating the supposed improvements in a fair
manner. There are two major challenges for such an immediate
validation. The rst one is methodological. Classical BPM lifecycle approaches
build on a labour-intensive analysis of the current process, which leads
to the deployment of a redesigned version [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. This new version is
monitored in production, and if it does not meet performance objectives, it is
made subject to analysis again. All this takes time. The second challenge
is architectural. Contemporary Business Process Management Systems
(BPMSs) enable quick deployment of process improvements, but they
do not provide support for validating improvement assumptions.
In this research, we address these challenges by integrating business
process execution concepts with ideas from modern software engineering
practices. We propose an iterative and incremental process improvement
methodology named AB-BPM that integrates business process execution
concepts with the idea of AB testing. AB-BPM supports the design of
AB tests with simulation and shadow tests.
      </p>
      <p>First, we develop a simulation technique that estimates the performance
of a new process version using historical data of the old version. Since the
results of simulation can be speculative, we propose shadow testing as
the next step. Our Shadow testing technique partially executes the new
version in production alongside the old version in such a way that the
new version does not throttle the old version. Finally, we develop
techniques for AB testing for redesigned processes with immediate feedback
at runtime. AB testing compares two versions of the deployed process
by observing users responses to versions A/B, and determines which one
performs better. We propose two algorithms, LTAvgR and
ProcessBandit, that dynamically adjust request allocation to the versions during the
test based on their performance. We evaluate these techniques with real
world and synthetic case studies.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Problem Statements</title>
      <p>Despite the prevalence of live testing and documentation of the successes,
contemporary business process management has not embraced this idea.
There are three problems with rolling out new process versions in
contemporary business process management practices:</p>
      <sec id="sec-2-1">
        <title>P1: New versions of business processes are not tested in production.</title>
        <p>The traditional BPM lifecycle encourages replacement of old versions
with new versions. The newly redesigned version makes the implicit
assumption that this version improves the business process. User
interactions are not observed before the replacement because the new version
was never executed in production. So, this improvement assumption is
not validated before committing to the new version.</p>
        <p>
          Live testing approaches are shown to have addressed such issues in the
context of websites (e.g. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]). However, such existing approaches do not
address complexities of BPM scenarios such as lengthy execution times
and involvement of human workers.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>P2: Evaluating new versions of a process in production is risky.</title>
        <p>
          There is an inherent risk of executing new versions that have not been
validated in production. We henceforth refer to this as risk of exposure.
Careful o ine analysis, as suggested by the BPM lifecycle, can reduce
the possibility of producing bad redesigns but the risk of exposure still
remains because a redesigned process completely replaces the old version.
Alternatively, live testing approaches follow the motto \test fairly, fail
fast", which means that there is minimal upfront analysis before testing
[
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. Typically, in techniques like AB Testing, the old and the new versions
are deployed simultaneously and half of the user requests are directed to
the new version. As such, the risk of exposing inferior versions is high.
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>P3: Process executions have to be evaluated with incomplete observations.</title>
        <p>To reduce risk of exposure, advanced AB testing techniques for websites
can observe metrics like click-through rate and dynamically adjust user
allocation during the test. This ensures that the better version receives
more tra c over time. However, such approaches cannot be used as-is
in BPM because of the complexities in performance indicators. Process
instances can have unique identi ers, and each execution can have
multiple process performance indicators such as duration and cost. Quality
of a process version can be assessed by observing performance indicators
of the corresponding process instances. First, the performance indicators
may be tied to duration. The delays in which these performance
indicators are observed in uences when and how the allocation of users to
the versions are adjusted. Second, not all of the required performance
indicators may be observable at the same time. For instance, duration
and user satisfaction scores may be obtained at di erent times with
delays of varying length. Furthermore, indicators like user satisfaction may
not be observed at all. Finally, the performance of a process can also
be in uenced by factors such as resource constraints, the environment,
and market uctuation. To adopt advanced AB testing techniques in
BPM, we require monitoring support, mechanisms of evaluating process
versions in presence of these complications with performance indicators,
and request allocation algorithms for the BPM scenario.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Contributions</title>
      <p>
        This research makes original contributions in the area of business
process improvement, business process simulation, and live testing. In this
research, we have proposed a business process improvement methodology
named AB-BPM [
        <xref ref-type="bibr" rid="ref6 ref8">6,8</xref>
        ]. It extends the business process lifecycle to provide
support for rapid and fair validation of process improvement hypotheses.
The AB-BPM methodology is supported by three classes of techniques:
simulation, shadow testing and AB testing. Fig. 1 illustrates how an old
version (A) is compared with a new version (B) using these techniques.
      </p>
      <sec id="sec-3-1">
        <title>Simulation</title>
        <p>
          Our simulation technique takes historical data (event logs) of a process
version, the process model of the new version, and estimates of the
metrics of some activities in the new version as the inputs. It produces event
logs for the new version, which is then used to estimate the performance
of the new version. We construct a Transition Simulation Tree (TST), a
rooted tree data structure that summarizes decisions and metrics
available in an event log of the existing process version.The TST created
from the old version is used in conjunction with a BPMS to derive the
execution log of the new version [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. We demonstrate the e cacy of this
approach with case studies from the domain of helicopter pilot licensing
and banking.
n
o
ilt
a
u
c
e
p
S
n
iltuao Version A
m
i
S
itgn User Request: rm
sT
e
oadw Copy of rm
h
S
        </p>
        <sec id="sec-3-1-1">
          <title>User Request: rm</title>
          <p>g
n
it
sT
e</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>AB User Request: rn</title>
        </sec>
        <sec id="sec-3-1-3">
          <title>AB-BPM Methodology</title>
        </sec>
        <sec id="sec-3-1-4">
          <title>Techniques</title>
        </sec>
        <sec id="sec-3-1-5">
          <title>Process Model</title>
          <p>(Version B)
+</p>
        </sec>
        <sec id="sec-3-1-6">
          <title>Event Logs</title>
          <p>(Version A)
+</p>
        </sec>
        <sec id="sec-3-1-7">
          <title>Metric</title>
        </sec>
        <sec id="sec-3-1-8">
          <title>Assumptions</title>
        </sec>
        <sec id="sec-3-1-9">
          <title>Version A</title>
          <p>(testVeenrvsiiroonnmBent)</p>
        </sec>
        <sec id="sec-3-1-10">
          <title>Version A</title>
        </sec>
        <sec id="sec-3-1-11">
          <title>Version B</title>
        </sec>
        <sec id="sec-3-1-12">
          <title>Performance Estimate of Version B</title>
        </sec>
        <sec id="sec-3-1-13">
          <title>Performance Estimate of Version B</title>
        </sec>
        <sec id="sec-3-1-14">
          <title>Best Version</title>
          <p>R
i
s
k
o
f
E
x
p
o
s
u
r
e
The risk of exposure is low in simulation because this technique is
performed o ine. However, the results can be speculative because of the
implicit assumptions about customer behavior and the explicit
assumptions about the execution of some tasks in the new version.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Shadow Testing</title>
        <p>
          Shadow testing executes the new version in production but hides the
execution from customers. User requests from the production environment
are duplicated and executed in both versions. Execution of the new
version can a ect the performance of the old version during test execution.
Therefore, the risk of exposure is higher than simulation. However, the
new version is executed online and the results are less speculative.
In our approach of shadow testing, we reduce the risk of exposure further
by partially simulating the new version and instantiating the new version
only when the performance overhead of testing is acceptable. As a result,
there is a trade-o between speculativeness and risk of exposure. This
is achieved through a new modular BPMS architecture, an overhead
management algorithm, and a decision tree for partial simulation [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
We evaluate this approach with synthetic and realistic processes from
the literature.
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>AB Testing</title>
        <p>
          AB testing executes the two versions in production such that each version
serves a portion of customers. This approach has high risk of exposure,
but the results of the tests are fair and reliable. To mitigate the risk of
exposure, we use multi-armed bandit algorithms that dynamically allocate
user requests to the process versions based on their performance. These
algorithms observe performance indicators of each version, and shift the
user requests towards the better performing version in the given context.
E ectively, the best version can be identi ed by the end of the test.
We propose a modular architecture that supports the execution of such
algorithms. We design two algorithms, LtAvgR and ProcessBandit, for
the BPMS case and propose reward designs for these algorithms [
          <xref ref-type="bibr" rid="ref6 ref7">6,7</xref>
          ]
Our AB testing solution observes performance of process execution in
the form of Process Performance Indicators (PPIs), convert the PPIs
into rewards, updates the rewards in case of incomplete observations (as
explained in P3), and dynamically allocates more requests to the version
that has better performance.
        </p>
        <p>We evaluate this solution with real-world case studies from the domain
of building permit approval, banking, and helicopter license approval.
We also demonstrate the e cacy of the algorithms with experiments on
convergence to optimal solution and scalability using synthetic cases.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>L.</given-names>
            <surname>Bass</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Weber</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhu. DevOps - A Software Architect</surname>
          </string-name>
          <article-title>'s Perspective. SEI series in software engineering</article-title>
          .
          <source>Addison-Wesley</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>M.</given-names>
            <surname>Dumas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. La</given-names>
            <surname>Rosa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mendling</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H. A.</given-names>
            <surname>Reijers</surname>
          </string-name>
          . Fundamentals of Business Process Management,
          <source>Second Edition</source>
          . Springer,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3. C. W. Holland.
          <article-title>Breakthrough Business Results With MVT: A Fast, Cost-Free \Secret Weapon" for Boosting Sales, Cutting Expenses, and Improving Any Business Process</article-title>
          . John Wiley &amp; Sons,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>K.</given-names>
            <surname>Kevic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Murphy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Williams</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and J.</given-names>
            <surname>Beckmann</surname>
          </string-name>
          .
          <article-title>Characterizing experimentation in continuous deployment: A case study on Bing</article-title>
          .
          <source>In International Conference on Software Engineering: Software Engineering in Practice Track</source>
          , pages
          <volume>123</volume>
          {
          <fpage>132</fpage>
          . IEEE Press,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>S.</given-names>
            <surname>Satyal</surname>
          </string-name>
          , I. Weber,
          <string-name>
            <given-names>H.</given-names>
            <surname>Paik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Ciccio</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Mendling</surname>
          </string-name>
          .
          <article-title>Shadow testing for business process improvement</article-title>
          .
          <source>In Confederated International Conferences: CoopIS</source>
          , pages
          <volume>153</volume>
          {
          <fpage>171</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>S.</given-names>
            <surname>Satyal</surname>
          </string-name>
          , I. Weber,
          <string-name>
            <given-names>H.</given-names>
            <surname>Paik</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Di Ciccio, and</article-title>
          <string-name>
            <surname>J. Mendling. ABBPM</surname>
          </string-name>
          <article-title>: performance-driven instance routing for business process improvement</article-title>
          .
          <source>In International Conference on Business Process Management, BPM</source>
          , pages
          <volume>113</volume>
          {
          <fpage>129</fpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>S.</given-names>
            <surname>Satyal</surname>
          </string-name>
          , I. Weber,
          <string-name>
            <given-names>H.</given-names>
            <surname>Paik</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Di Ciccio, and</article-title>
          <string-name>
            <surname>J. Mendling. AB</surname>
          </string-name>
          <article-title>Testing for process versions with contextual multi-armed bandit algorithms</article-title>
          .
          <source>In International Conference on Advanced Information Systems Engineering (CAiSE)</source>
          , pages
          <fpage>19</fpage>
          {
          <fpage>34</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>S.</given-names>
            <surname>Satyal</surname>
          </string-name>
          , I. Weber,
          <string-name>
            <given-names>H.</given-names>
            <surname>Paik</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Di Ciccio, and</article-title>
          <string-name>
            <given-names>J.</given-names>
            <surname>Mendling</surname>
          </string-name>
          .
          <article-title>Business process improvement with the AB-BPM methodology</article-title>
          .
          <source>Information Systems</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>