=Paper= {{Paper |id=Vol-1674/iStar16_pp55-60 |storemode=property |title=An Empirical Evaluation Roadmap for iStar 2.0 |pdfUrl=https://ceur-ws.org/Vol-1674/iStar16_pp55-60.pdf |volume=Vol-1674 |authors=Lidia López,Fatma Başak Aydemir,Fabiano Dalpiaz,Jennifer Horkoff |dblpUrl=https://dblp.org/rec/conf/istar/LopezADH16 }} ==An Empirical Evaluation Roadmap for iStar 2.0== https://ceur-ws.org/Vol-1674/iStar16_pp55-60.pdf
    An Empirical Evaluation Roadmap for iStar 2.0

Lidia López1 , Fatma Başak Aydemir2 , Fabiano Dalpiaz2 , and Jennifer Horkoff3
    1
            Universidad Politècnica de Catalunya, Barcelona, Spain, llopez@essi.upc.edu
        2
             Utrecht University, Utrecht, Netherlands, {f.b.aydemir, f.dalpiaz}@uu.nl
                   3
                     City University London, London, UK, horkoff@city.ac.uk


              Abstract. The iStar 2.0 modeling language is the result of a two-year
              community effort intended at providing a solid, unified basis for teaching
              and conducting research with i*. The language was released with impor-
              tant qualities in mind, such as keeping a core set of primitives, providing
              a clear meaning for those primitives, and flattening the learning curve
              for new users. In this paper, we propose a list of qualities against which
              we intend iStar 2.0 to be evaluated. Furthermore, we describe an em-
              pirical evaluation plan, which we devise in order to assess the extent
              to which the language meets the identified qualities and to inform the
              development of further versions of the language. Besides explaining the
              objectives and steps of our planned empirical studies, we make a call for
              involving the research community in our endeavor.
              Keywords: i* Framework, iStar 2.0, empirical engineering, evaluation


1           Introduction
Many dialects and extensions of the i * modeling language have been proposed
since its introduction in the 1990s. Although these proposals demonstrate the
popularity of the language in the research community and allow adaptation of
the framework to a variety of domains (e.g., security, law, service-oriented archi-
tectures), they have also created difficulties in learning, teaching, and applying
i * consistently.
     iStar 2.0 [2] is the result of a collective effort of the i * community aimed to
overcome these difficulties by defining a standard core set of concepts. Given the
objectives of iStar 2.0, our aim is twofold: a. to measure how well the language
achieves the objectives, and b. to inform further developments with empirical
evidence.
     More specifically, our research question is the following: Does iStar 2.0 pro-
vide a solid and unified basis for teaching and supporting ongoing research on
goal-oriented requirements engineering? Toward answering this question, we
identify several relevant qualities and provide an initial roadmap for the em-
pirical studies to conduct to evaluate iStar 2.0 against those qualities.
     The remainder of the paper is structured as follows. Section 2 includes a
brief literature review of empirical evaluations in modeling languages and i *. In
Section 3, we define the set of qualities to be empirically evaluated and a roadmap
defining the time-line of the implementation of these evaluations. Finally, we
draw some conclusions in Section 4.

 Copyright © 2016 for this paper by its authors. Copying permitted for private and academic purposes.
Proceedings of the Ninth International i* Workshop (iStar 2016), CEUR Vol-1674

2      Empirical Evaluation of Modeling Languages
There is a variety of empirical evaluations in the area of modeling languages in
general, and in i * modeling language in particular. This section provides a brief
summary of these studies focusing on the qualities evaluated by the studies.
    Lindland et al. [7] propose a framework that identifies three category of
qualities related to modeling languages (syntactic, semantic, and pragmatic),
quality goals for each category, and the means for achieving these goals. The
semantic qualities refer to the validity and completeness of the language and
the models generated using the language, syntactic qualities are related to the
syntax of the language, and pragmatic qualities concern the understandability
of the language and its application.
    Guizzardi et al. [5] suggest domain and comprehensibility appropriateness as
key qualities of a modeling language, relying on verifying lucidity, soundness,
laconicity, and completeness properties of model instances. These properties are
then related to corresponding language properties: construct overload, construct
excess, construct redundancy, and ontological expressiveness.
    Frank [4] proposes a method to evaluate reference models, where the evalu-
ation concerns both the general qualities of conceptual models and re-usability
of the reference domain. The framework states four different evaluation perspec-
tives: economic, deployment, engineering and epistemological. Each perspective
is structured into multiple aspects for which a success criterion is provided.
    Interest in i * evaluation appears to be on the rise, with studies covering
both the language evaluation and the applicability of i * in the industry. We
distinguish between different kinds of studies. Some works evaluate the use of
an i * extension comparing it to the use of i * [10]. Other approaches compare
i * with other goal-oriented modeling languages such as KAOS [8] or Techne [6].
Finally, other studies evaluate specific characteristics of the language such as
visual effectiveness [9].
    The majority of the studies providing empirical evidence in the literature
are evaluating the applicability of i * for different purposes in the industrial
environment. Elahi et al. [3] studied the use of i * for gathering and understanding
knowledge in an organization, concluding that some constructs are not used by
practitioners. Carvallo et al. [1] focus on socio-technical systems and conclude
that some models are too difficult to read and modify due to their complexity.
A variety of real use cases were presented at the i * Showcase in 2011 1 .


3      iStar 2.0 Evaluation Roadmap
In order to evaluate iStar 2.0, we need to define the set of the language qualities
that we want to assess. Based on the review of Section 2, we present a number of
qualities to evaluate, then discuss suitable empirical methods, and finally devise
an initial roadmap for the empirical evaluation.
 1
     http://www.city.ac.uk/centre-for-human-computer-interaction-design/
     istar11


                                            56
                                         An Empirical Evaluation Roadmap for iStar 2.0

3.1   Qualities to be evaluated
As iStar 2.0 was not defined as a new language, but a set of core concepts
refining the original i* [12], backward compatibility is critically important. As
a community, we need to collect evidence to determine if iStar 2.0 meets the
needs of the users of i *. The open nature of i * comes with a drawback that iStar
2.0 is trying to mitigate: the steep learning curve that makes it hard to employ
the language in industry. Therefore, learnability is also a priority quality to be
evaluated. Keeping the open nature of i * was also one of the main objectives
during the definition of iStar 2.0. Consequently, we also need to consider the
extensibility quality, i.e., evaluating whether iStar 2.0 is a suitable baseline for
extensions.
    Additionally to these qualities, we consider some qualities to evaluate the
quality of the language, for example expressiveness or syntactic correctness. Re-
garding the expressiveness, we are interested in evaluating if iStar 2.0 has a
suitable set of constructs (missing, excess or overload). Syntactic correctness
evaluates whether the modelers can easily detect and correct syntactic errors
using iStar 2.0 and whether the language can prevent syntactic errors.
    We have also included qualities not directly assessed during the definition
of iStar 2.0, such as scalability. The detailed set of qualities to be evaluated
is included in Table 1. We categorize the qualities based on the classification
provided in [7].

3.2   Empirical Methods: Design Dimensions
In order to evaluate the qualities listed in Table 1, several empirical studies must
be designed and conducted. We envision the application of several empirical
methods, including experiments, surveys and case studies. We can enumerate a
number of dimensions that must be considered when designing such studies.
    The choice of subjects participating in the studies is a dimension that must
be determined for each study. To classify the subjects, we can use two categories:
expertise and background (industry or academy). We need to clearly define a set
of i* experts for inclusion in the backward compatibility evaluation. For practical
applicability, we need to involve practitioners from industry. For other qualities,
we can treat the expertise and the background of participants as a variable in
the study.
    We also need to decide when to evaluate the iStar 2.0 language in isolation
and when a comparative analysis comparing iStar 2.0 to i * is needed. The same
reasons that lead us to pay special attention to the backward compatibility and
the learnability lead us to think, that for these specific qualities, we should
conduct comparative analysis. Meanwhile, the evaluation of the other qualities
can focus only on iStar 2.0.

3.3   Roadmap
From an empirical software engineering standpoint, we can identify two main
phases for the evaluation of iStar 2.0: formative and summative. The formative


                                        57
Proceedings of the Ninth International i* Workshop (iStar 2016), CEUR Vol-1674


                     Table 1: iStar 2.0 qualities to be evaluated
  Category      Quality              Definition
                                     Does iStar 2.0 facilitate ensuring and maintaining
  Syntactic     Syntactic correct-
                                     syntactic correctness?
                ness
                                     Does iStar allow one to capture a sufficient num-
                Expressiveness
                                     ber of concepts in a socio-technical domain?
  Semantic
                Unambiguous          Do iStar 2.0 models have only one interpretation?
                models
                                     Is iStar 2.0 able to represent the same phenomena
                Backward–
                                     as i*?
                ccompatibility
                Comprehensibility Can iStar 2.0 models be understood?
                                   Is the effort required to use iStar 2.0 worth the
                Cost–
                                   benefits?
                Effectiveness
  Pragmatic     Extensibility      Is it easy to add new concepts to iStar 2.0?
                                   What does the learning curve of iStar 2.0 look
                Learnability
                                   like?
                                   Does iStar 2.0 facilitate changing and updating
                Modifiability
                                   models?
                                   Can iStar2.0 be successfully applied to real world
                Practical applica-
                                   cases?
                bility
                                   Does iStar 2.0 support the creation and analysis
                Scalable
                                   of large problems?




phase corresponds to the task related to development of the proposal providing
some partial empirical validation for the resulting proposal, while the summative
phase evaluates if the proposal can be implemented in the real world. We are
currently in the formative phase, and precisely in the treatment validation step
of Wieringa’s design science methodology [11].
    We divide the proposed empirical evaluation plan in three phases, divided
in a total of five stages. The first two phases correspond to the formative and
summative phases in empirical research, while the third phase describes comple-
mentary activities:
  – In the formative phase, the evaluation will concern the qualities that led the
    design decisions for iStar 2.0. These qualities include keeping a core set of
    primitives (stage 1), providing a clear meaning for such primitives (stage 2),
    and flattening the learning curve for new users (stage 3).
  – In the summative phase, the proposal (in our case, iStar 2.0) should be tested
    for applicability to real-world cases (stage 4).
  – The third phase includes the study of additional properties that do not
    directly relate to the use of iStar 2.0 itself, but rather on its capability to be
    adapted for specific cases or domains (stage 5).


                                            58
                                          An Empirical Evaluation Roadmap for iStar 2.0

Figure 1 shows the three phases, including the qualities to be evaluated in each
stage. Cost-effectiveness is a quality that should be evaluated as part of all the
stages. The cost can be evaluated in terms of time in all the stages, and also in
terms of money in stage 4. Note that stages 1 to 3 and 5 could be executed in any
order while stage 4 should be executed after stages 1 to 3 have been conducted.




                      Fig. 1: iStar 2.0 Evaluation Roadmap




4   Conclusions
During the last few years, the i * community has been working on the definition
of a standard, core version, resulting in iStar 2.0. The main goal of this effort
was to facilitate the consistent learning, teaching, and application of i *. After
the definition of iStar 2.0, the natural next step is evaluating the proposal to
gather evidence of whether the proposal achieves the expected qualities.
    In this paper, we emphasize the necessity of evaluating iStar 2.0 through
empirical studies. Our first step is the identification of a set of qualities against
which iStar 2.0 should be evaluated. We also discuss some key dimensions that
need to be defined when conducting these empirical studies. Interestingly, many
of the qualities we identified are pragmatic; we surmise this is linked to the
limited adoption of i * in industry.
    We prioritize the evaluation tasks of the qualities grouping them in five stages.
Some of these tasks are labeled as formative evaluation, others are part of sum-
mative evaluation, and the remaining tasks are additional studies of the exten-
sibility and customizability of iStar 2.0. Based on this grouping, we define a
roadmap proposing an order of execution for the various evaluation stages.
    The next steps consist of conducting empirical studies addressing one or more
of the identified qualities for iStar 2.0. Although we plan to design and conduct
several studies ourselves, an effective evaluation of the language will require a
community-wide effort. We encourage i * community members to use and eval-
uate iStar 2.0, keeping in mind the qualities presented here, and reporting the


                                        59
Proceedings of the Ninth International i* Workshop (iStar 2016), CEUR Vol-1674

results publicly. Our hope is that, as a community, we build evidence either to
support the usefulness of iStar 2.0 as well as to shape the future versions of the
language.
Acknowledgments. This work is supported by EOSSAC project, funded by the
Ministry of Economy and Competitiveness of the Spanish government (TIN2013-
44641-P), an ERC Marie Skodowska-Curie Intra European Fellowship (PIEF-
GA-2013-627489) and a Natural Sciences and Engineering Research Council of
Canada Postdoctoral Fellowship (Sept. 2014 – Aug. 2016). The second and third
author have received funding from the SESAR Joint Undertaking under grant
agreement No. 699306 under European Union’s Horizon 2020 research and in-
novation programme.


References
 1. Carvallo, J.P., and Franch, X.: “On the Use of i* for Architecting Hybrid Systems:
    A Method and an Evaluation Report”. In: Lecture Notes in Business Information
    Processing. Springer Science + Business Media, 2009, pp. 38–53.
 2. Dalpiaz, F., Franch, X., and Horkoff, J.: iStar 2.0 Language Guide. CoRR abs/
    1605.07767 (2016)
 3. Elahi, G., Yu, E., and Annosi, M.C.: “Modeling Knowledge Transfer in a Soft-
    ware Maintenance Organization - An Experience Report and Critical Analysis”.
    In: Lecture Notes in Business Information Processing. 2008, pp. 15–29.
 4. Frank, U.: “Evaluation of Reference Models”. In: Reference modeling for business
    systems analysis. IGI Global, 2006, pp. 118–140.
 5. Guizzardi, G., Pires, L.F., and Sinderen, M. van: “An Ontology-Based Approach for
    Evaluating the Domain Appropriateness and Comprehensibility Appropriateness
    of Modeling Languages”. In: Model Driven Engineering Languages and Systems.
    Springer Science + Business Media, 2005, pp. 691–705.
 6. Horkoff, J., Aydemir, F.B., Li, F., Li, T., and Mylopoulos, J.: Evaluating Modeling
    Languages: An Example from the Requirements Domain. In: ER (2014)
 7. Lindland, O., Sindre, G., and Solvberg, A.: Understanding Quality in Conceptual
    Modeling. IEEE Softw. 11(2), 42–49 (1994)
 8. Matulevičius, R., and Heymans, P.: “Comparing Goal Modelling Languages: An Ex-
    periment”. In: Requirements Engineering: Foundation for Software Quality. Springer
    Science + Business Media, 2007, pp. 18–32.
 9. Moody, D.L., Heymans, P., and Matulevicius, R.: Improving the Effectiveness of
    Visual Representations in Requirements Engineering: An Evaluation of i* Visual
    Syntax. In: RE (2009)
10. Teruel, M.A., Navarro, E., López-Jaquero, V., Montero, F., and González, P.:
    “Comparing Goal-Oriented Approaches to Model Requirements for CSCW”. In:
    Communications in Computer and Information Science. 2013, pp. 169–184.
11. Wieringa, R.J.: Design science methodology for information systems and software
    engineering. Springer (2014)
12. Yu, E.: Modelling strategic relationships for process reengineering. University of
    Toronto (1996).




                                            60