=Paper=
{{Paper
|id=Vol-1442/paper_5
|storemode=property
|title=An Empirical Study to Validate the Use of Ontological Guidelines in the Creation of i* Models
|pdfUrl=https://ceur-ws.org/Vol-1442/paper_5.pdf
|volume=Vol-1442
|dblpUrl=https://dblp.org/rec/conf/ontobras/JuniorGFGW15
}}
==An Empirical Study to Validate the Use of Ontological Guidelines in the Creation of i* Models==
An Empirical Study to validate the Use of Ontological Guidelines in the Creation of i* Models Ramilton Costa Gomes Júnior1, Renata Silva Souza Guizzardi1, Xavier Franch2, Giancarlo Guizzardi1, Roel Wieringa3 1 Ontology and Conceptual Modeling Research Group (NEMO). Federal University of Espírito Santo, Vitória/ES, Brasil. 2 Universitat Politècnica de Catalunya (UPC) Barcelona, Spain University of Twente 3 Enschede, The Netherlands ramiltoncosta@gmail.com,{rguizzardi,gguizzardi}@inf.ufes.br, franch@essi.upc.edu,roelw@cs.utwente.nl Abstract. i* is a well known goal modeling framework, developed by a large and geographically dispersed research community. Currently, i* users tend to ascribe different and conflicting meanings to its constructs, leading to a non uniform use of the language, and consequently undermining its adoption. In previous works, we proposed ontological guidelines to support the creation of i* models, in an attempt to provide a solution to this problem. In this paper, we present an empirical study, to evaluate these ontological guidelines. Results show that for more experienced conceptual modelers, the ontological guidelines indeed support i* modeling. However, results are not as positive for nonexperienced conceptual modelers. 1. Introduction i* is a goal modeling framework used for Requirements Engineering (Yu, 1995). In the past twenty years, this framework has attracted the attention of different research groups, which haveproposed different variants of the initial framework, each one proposing different semantics to the language's constructs. The community that develops i* is aware that this nonuniform use of i* makes it difficult for novices to learn how to use the language, besides undermining its acceptance in industry. We believe this problem can be solved with the use of a foundational ontology to interpret the semantics of the i* concepts. A foundational ontology is a formal system of domainindependent categories that can be used to characterize the most general aspects of concepts and entities that belong to different domains in reality (Guizzardi, 2005). The idea is to apply the foundational ontology as a reference model to interpret the concepts of the language. Then, based on such interpretation, we are able to provide some guidelines to support modeling, here referred as ontological guidelines. In previous works (Guizzardi, Franch, Guizzardi, 2012), (Guizzardi, Franch, Guizzardi, Wieringa, 2013), we proposed some ontological guidelines for i* modeling, based on the UFO foundational ontology (Guizzardi, 2005) (Guizzardi et al, 2013)(Guizzardi, Falbo, Guizzardi, 2008). The aim of this paper is to present the experimental design and the results of an empirical study conducted to evaluate the use of such ontological guidelines. Nowadays, empirical studies are considered appropriate means to prove the effectiveness of a new approach. For (Vokac, 2002), the ideal science would have a set of empirical observations for each theory, either to support the theory or to prove it wrong. In other words, empirical observation is the core of the scientific process. Furthermore, it is through empirical observation that one can check theories, explore critical factors and give light to new phenomena, so that the theories can evolve (Travassos, 2002). Having this in mind, we decided to conduct an experiment to confirm our intuitions that the use of ontological guidelines lead to i* models of better quality. The experiment was conducted in two colleges, having fiftyfive subjects in total. The subjects were students of a Systems Analysis and Development course and the PhD and Master program in Computer Science. The main goal of the experiment was to verify if the ontological guidelines cited above are useful or not in the development of i* models. For that, the subjects participated in modeling activities with and without the use of the guidelines and, then, the results were compared. In the experiment applied with PhD and master students, the results show that the ontological guidelines are useful for the development of i* models. Among the population of the second experiment application, composed of less experienced conceptual modelers, the experiment results were not so positive. The remainder of this article is organized as follows: Section 2 presents information on the i* framework and its variants; Section 3 describes the UFO fragment applied in this work; Section 4 presents some of the proposed ontological guidelines; Section 5 describes the empirical study; and, finally, Section 6 concludes the paper. 2. The i* Framework and its Variants The original i* framework is described in (Yu, 1995). Since then, several variants have been proposed, for instance GRL and Tropos, see (Cares, 2012) for an overview. Some variants come from paradigm shifts, others propose some particular type of new construct, and still others issue slight modifications related to the core constructs of the i* language. One of the most controversial constructs in the language is the meansend link. In the original i* (Yu, 1995), this link is used to connect a goal or a task to softgoals. In GRL, this link is applied to connect a task to a goal, a task to a task and a resource to a task. However, in the i* wiki, one of the major sources of material about the language, this link is only used to connect a task to a goal. (Cares, 2012) also points out that different versions of Tropos propose different uses for the meansend link. These different interpretations and uses make a new i* learner confuse. She may ask herself: when can I use a meansend link after all? Why is it used this way? Why can't I use a meansend link between a resource and a goal, for example? We argue that the best way to respond to these questions is to understand the ontological semantics behind the constructs of the language. By understanding their ontological nature, we may provide good reasons why a concept or a link may or may not be used in a particular way. 3. Background: The UFO Foundational Ontology Here we briefly present the UFO concepts that are used in this paper provide an interpretation to i*. To facilitate reading we use a different font to highlight the UFO concepts. For a fuller presentation on UFO, the reader should refer to (Guizzardi, 2005), (Guizzardi et al, 2013) and (Guizzardi, Falbo, Guizzardi, 2008). In UFO, a stakeholder is represented by the Agent concept, defined as a concrete Endurant (i.e. an entity that endures in time while maintaining its identity) which can bear certain Intentional States. These intentional states include Beliefs, Desires and Intentions. Intentions are mental states of Agents which refer to (are about) certain Situations in reality. Situation are snapshots of reality. The propositional-content (i.e., proposition) of an Intention is termed a Goal. In contrast to Endurants, Events are perduring entities, i.e., entities that occur in time accumulating their temporal parts. Events are triggeredby certain Situations in reality (termed their presituations) and they change the world by producing a different postsituation. Action are deliberate Events, i.e., Events deliberately performed by Agents in order to fulfill their Intentions. An Action achieves a Goal if the Action brings about a Situation in the world that satisfies that Goal. In contrast with an Agent, an Object is a concrete Endurant that does not bear intentional states or perform actions. An Object participating in an Action is termed a Resource. 4. Ontological Guidelines for the Creation of i* Models. In this section, we describe some of the proposed ontological guidelines. For lack of space, we are not able to present them all and refer to (Guizzardi, Franch, Guizzardi, 2012) and (Guizzardi, Franch, Guizzardi, Wieringa, 2013) for a full description. In total, there are seven ontological guidelines and all of them have been considered in the experiment. First, it is important to point out that we interpret i* goals, tasks, resources and agents as their counterparts in UFO (with Action as task). Having that in mind, let us try to interpret the i* decomposition relation. Since goals are propositions, due to its ontological nature, it is impossible for a goal to be decomposed into tasks or resources. Thus, goals can only be decomposed into subgoals. Consequently, when decomposing goals, an i* anddecomposition is interpreted as a conjunction of subgoals, while an i* ordecomposition is interpreted as a disjunction of subgoals. Similarly, softgoals, tasks and resources can only be decomposed into softgoal, tasks and resources, respectively. This originates the ontological guideline describe in the first line of Table 1. In i*, a meansend link is applied to connect a means to an end. For example, a task T (means) to a goal G (end), meaning that the execution of T leads to the achievement of G. Here, we adopt the conceptual modeling evaluation method proposed in (Guizzardi, 2005) that states that we should avoid construct redundancy, i.e., two language constructs should not be applied to model the same phenomenon in the world. Construct redundancy adds unnecessary complexity to the modeling language, besides making specifications more difficult to understand. Moreover, when facing redundancy, designers tend to ascribe slightly different meanings to the redundant constructs, which may not be fully understood by the model readers. In our case, if we allow, for instance, goals G2 and G3 to be connected via meansend to goal G1, we will not be able to differentiate between meansend and ordecomposition, i.e. these two links will be applied to represent the very same relation in the world. Thus, this will be a case of construct redundancy. To avoid that, we propose the ontological guideline described in the second line of Table 1. In i*, a makecontribution is applied between a task T and a goal G, meaning that if T is executed, then G is fully achieved. But if this is so, how can one differentiate between meansend and makecontribution? Using UFO, we differentiate this by looking at the intention behind the execution of T. To understand this, let us consider the i* model depicted in Figure 1, which exemplifies the use of the meansend and the make contribution links. Figure 1. Meansend vs. makecontribution In Figure 1, a Car Passenger1 agent executes the Take a car sick pill task in order to prevent himself from being sick during the journey he is making (meansend link to Car sickness prevented goal). As a side effect of this medication, the Car Passenger also goes to sleep (makecontribution link to Asleep fallen goal). As result of the mapping from i* tasks into UFO actions, every task is associated with a motivating intention whose propositional content is a goal. In other words, we 1 From now on, we use a different font for the names of the instances of the i* actors and intentional concepts, such as goals, tasks, and resources. execute a particular task in order to accomplish a specific goal. In i*, the association between the task and the goal in this case is made by a meansend link (e.g. Take a car sick pill task as means to Car sickness prevented goal). On the other hand, this same task can also generate some other goals to be accomplished, without however, being intended be the choice of this particular task. In this case, a makecontribution link is established (e.g. Take a car sick pill task contributing to asleep fallen goal). In other words, the meansend link or the makecontribution link should be applied according to the ontological guideline described in the third line of Table1. Table 1. Some of the proposed i* ontological guidelines Ontological Guidelines 1. A decomposition link can only be applied between elements of the same kind. E.g. goal>goal, task>task. 2. A meansend link can only be applied between elements of different kinds. E.g. task>goal, resource>task. 3. Taking task T and goal G, if the intention behind the execution of task T is to accom plish G, T and G should be related via meansend link. On the other hand, if by execut ing T, G is unintentionally achieved (i.e., as a sideeffect of the execution of T), then T and G should be related via makecontribution. 5. The Empirical Study In this section, we describe the empirical study we conducted to evaluate the use of the ontological guidelines. The hypothesis of the study is "the ontological guidelines enhance the capability of the subjects to create i* models." The experiment was conducted in a controlled environment and is based on a quantitative strategy, in which the data is analyzed using statistical and descriptive methods. For the experimental design, we followed the framework presented in (Kochanski, 2009). 5.1 Experimental Design The experiment has as object of study two i* models (here referred to as Case 1 and Case 2), representing two different situations. Each participant had to complete the models, by filling in the blanks with the correct element or link to be used in each question. Figure 2 illustrates part of one model. For each blank, there are two and more possibilities, having as alternatives constructs of i* whose use normally generates confusion or doubts. For example, in Question 2 (refer to Figure 2), the participants should indicate if “Provide gift wrapping solution” is a goal or a plan. In Question 5, the participants should indicate if “Provide gift wrapping solution” and the two tasks “Organize wrapping stand” and “Allow vendors to wrap gifts” should be linked via OR meansend or via ORdecomposition. The idea is to verify if the participants can select them intuitively (pretest) or if the use of ontological guidelines (posttest) effectively helps the selection of the correct construct. Figure 2. Part of the i* model The experiment was divided in two steps: pretest and posttest. In the pretest, all participants performed the first activity, i.e. filling in the blanks, using Case 1. Then, in a separate form, they justified their choices for each blank. During this activity, all participants had a printout of some slides containing basic information about i* (the i* wiki guidelines), as well as the description of Case 1. No information about the guidelines is given in this first step. After the pretest activity, the students were randomly divided into two groups: group A (control group) and group B (experimental group). After the division, the participants of group A moved to another room to perform the posttest activity. Both groups had to perform a second activity of filling in the blanks, now using Case 2. However, in this part, only group B received information about the ontological guidelines. Both groups had the description of Case 2 and group B also had a printout of some slides containing the ontological guidelines. In the posttest, the participants of both groups were also asked to fill in a separate form justifying their choices for each blank. To capture the impression of the participants about the guidelines, the participants were also asked to respond some questions regarding their opinion about the i* wiki guidelines and the ontological guidelines. 5.2 Collected Data The data was collected through questionnaires. Before the experiment activities, we applied a questionnaire to capture the participants’ profile. We applied the experiment twice, with two different populations. We will here refer to these applications as application 1 and application 2. In application 1, there were 24 participants: 16 of them were undergraduate students of Computer Science or Computer Engineering, 7 of them were master students in Computer Science, and 1 of them was a PhD student in Computer Science. The participants were assigned into two groups of 12 participants, which were balanced in terms of educational level and modeling experience. In both groups, there was one participant with 13 years of experience in goal modeling and i*, while the others declared not having experience in this area. In application 2, there were 30 participants, all of them in the final year of an undergraduate course in Information Systems Analysis and Development. Each group had 15 participants. None of the participants indicated having experience in goal modeling or i*. Both in the pretest and in the posttest, the same activities and questionnaires were used in applications 1 and 2. The graphs of Figures 3 and 4 show the results for the first and the second application, respectively. When the participant fills in the blank correctly, we say that he has a hit. Figure 3. Hits by participant in pretest (left) and posttest (right) in the first experiment application. Figure 4. Hits by participant in pretest (left) and posttest (right) the second experiment application. Tables 2 and 3 present data regarding the number of hits per participant in the first and second application, respectively. The columns present data on average, median, highest and lowest value of number hits per participants. Table 2 Number of hits per participants in the first application Average Median Highest Lowest Group A Group B Group A Group B Group A Group B Group A Group B Pretest 6,67 5,50 5,50 5,00 8,00 8,00 4,00 3,00 Posttest 9,00 11,00 9,00 11,50 11,00 13,00 7,00 8,00 Table 3 Number of hits per participants in the second application Average Median Highest Lowest Group A Group B Group A Group B Group A Group B Group A Group B Pretest 5,87 6,20 6,00 6,00 5,00 10,00 2,00 3,00 Posttest 7,89 9,27 9,27 8,00 10,00 13,00 4,00 5,00 5.3 Data Analysis Analyzing Figure 3, we notice that in the pretest of the first application, the participants of group A scored a larger number of hits than the participants of group B. However, in the posttest, group B performed better than group A. This shows that the group that used the ontological guidelines performed better when compared to the group that only had access to the i* wiki guidelines. This result favors our hypothesis, supporting the idea that the ontological guidelines effectively help the creation of i* models. By looking at Figure 4, we see that in the pretest of the second application, groups A and B showed a great balance in realizing the activities; both groups scored the same number of hits and errors. In the posttest, group B achieved a significantly higher number of hits in relation to group A, as seen in Figure 4. Again, this result favors our hypothesis, supporting the idea that the ontological guidelines effectively help the creation of i* models. Table 2 shows the data regarding the number of hits per participants in the pre test and posttest, in the first application. The values for average, median, highest and lowest are very similar in the pretest activity. But in the posttest activity, the values are significantly different, result that favors ours hypothesis. Table 3 presents the data regarding number of hits per participants in the pretest and posttest, in the second application. The values for average, median, highest and lowest have a small difference in the pretest activity. But in the posttest activity, the values are significantly different, result that favors ours hypothesis. The descriptive analysis we presented so far is able to provide us with some evidence supporting the hypothesis, We can quantify this support by a statistical test. Thus, we also applied the WilcoxonMannWhitney statistical test, with a significance level of 5%, to compare the hits for each participant between the experimental (group B) and control (group A) groups, in both experiment applications. This statistical method is a nonparametric method recommended for small samples or groups with less than 20 participants (Robson 2002). In the first application, the calculated U value is 23 and the critical U value from the MannWhitney index is 37. Since the calculated U is lower than the critical U, then we may conclude that the values are significantly different between the groups, which supports our hypothesis. In the second application, the calculated U value is 65 and the critical U value from the MannWhitney index is 64. Since in this case, the calculated U is not lower than the critical U, then we cannot confirm our hypothesis. Given the results of the MannWhitney test, we cannot conclude that the ontological guidelines are always helpful. We attribute this difference to the divergence in profiles in the two experiment applications. The participants of the first application have a higher graduation level than the participants of the second application, and thus are, in general, more experienced in conceptual modeling. Thus, we claim that the ontological guidelines are helpful for more mature conceptual modelers. New empirical studies should be conducted to confirm this hypothesis. Regarding the qualitative evaluation of the ontological guidelines, we have the following results. In the first application, 7 out of 12 participants considered that the ontological guidelines are better than the i* wiki guidelines. The other 5 participants considered that the ontological guidelines and the i* wiki guidelines have the same quality. When asked about the usefulness of the ontological guidelines, 8 participants considered them very useful, 2 participants found them not very useful and 2 participants found them indifferent. In the second application, 13 out of 15 participants considered that the ontological guidelines are better than the i* wiki guidelines, while 2 participants considered that the ontological guidelines and the i* wiki guidelines have the same quality. Regarding the usefulness of the ontological guidelines, 10 participants found them very useful, 3 participants found them not so useful and 2 found them indifferent. We find these results positive, as most of the participants had a good perception regarding the ontological guidelines. Let us now analyze which questions were more difficult, i.e. led to more errors in both experiment applications. This will allow us to find out which ontological guidelines are not clear and should be improved. In the first application, the questions that led to more errors were questions 8 and 10. In the second application, the questions that led to more errors were questions 7, 9 and 14. Questions 8, 9, 10 and 14 regard the use of the meansend, makecontribution and helpcontribution links. We conclude that the participants in both experiment applications could not understand well the ontological difference between these three links. Thus, the ontological guidelines concerning this differentiation should be improved. Question 7 regards the differentiation among AND and OR decomposition. We conclude that in the second application, the participants also had doubts regarding the use of decomposition. Thus, the guidelines concerning these links should also be improved. 5.4 Threats to Validity The following factors are considered the main threats to the validity of this empirical study: a) the heterogeneity of the participants of the first application, since they had different academics degrees. To mitigate this risk, we collected information about the academic degree of the participants in the profile questionnaire and took this into account in our experiment design; b) the possibility that the participants had previous knowledge of the ontological guidelines. To remediate this risk, we asked in the experiment questionnaire if the participant had had previous contact with the guidelines. This information was taken into account in our analysis; c) the chance that the participants had low interest in the experiment results, carelessly performing the experiment activities. To mitigate this risk, we tried to motivate the participants, showing the importance of the results of the experiment. Moreover, the experiment was designed to be as short as possible, so as to prevent tiredness and disinterest; d) the possibility that the researcher conducting the experiment influenced the experiment results. To remediate this risk, the researcher conducting the experiment tried to be as objective and unbiased as possible during the experiment activities; e) the possibility that the subjects had a positive opinion about the guidelines, because they knew we were the ones who formulated them. To remediate this risk, we did not tell them we were the authors of the guidelines. 6. Final Considerations This article presented an empirical study with the objective to evaluate the use of ontological guidelines to create i* models. For that, the experiment was conducted in two steps (pre and posttest), in which the participants performed modeling activities without (pretest) and with (posttest) the use of ontological guidelines. To analyze the results, we performed the MannWhitney statistical test. The outcome supports our hypothesis that states that the guidelines are useful, and does not provide evidence against it. Moreover, most participants stated that they found the ontological guidelines useful to support them in the creation of i* models. Given the results of this experiment, we intend to develop an i* modeling tool that uses the ontological guidelines as support for the model designer. For that, we aim at proposing a metamodel that is compatible with these guidelines, to serve as basis for the development of the tool. For the future, we also intend to perform new experiments to collect more data regarding the use of the ontological guidelines to create i* models. In order to confirm our hypothesis, we must repeat the designed experiment, taking populations of different profiles. We aim, for example, to conduct the experiment with professional modelers. Moreover, we intend to perform different experiments. For instance, we would like to conduct an experiment in which the participants are asked to create i* models from scratch, with and without the use of the ontological guidelines. Then, based on some preestablished criteria collected from i* experts, we will be able to analyze if the models created with the use of ontological guidelines have higher quality than the ones created without them. Acknowledgement. This work is partially supported by CAPES/CNPq (grant number 402991/20125), CNPq (grant numbers 461777/20142 and 485368/20137), and the Spanish project EOSSAC, ref. TIN201344641P. References Ayala, C., Cares, C., Carvallo, J.P., Grau, G., Haya, M., Salazar, G., Franch, X., Mayol, E. and Quer, C. (2005), “A Comparative Analysis of i*Based AgentOriented Modeling Languages”, In: 17th International Conference on Software Engineering and Knowledge Engineering, Taipei, Taiwan, pp. 4350. Cares, C. (2012), “From the i* Diversity to a common interoperability framework”, PhD Thesis, Software Engineering for Information System Research Group, UPC, Spain. Guizzardi, G. (2005), “Ontological Foundations for Structural Conceptual Models”. PhD Thesis, University of Twente, The Netherlands. Guizzardi, G., Wagner, G., Falbo, R.A., Guizzardi, R.S.S., Almeida, J.P.A. (2013), Towards Ontological Foundations for the Conceptual Modeling of Events, 32nd International Conference on Conceptual Modeling (ER 2013), Hong Kong. Guizzardi, G., Falbo, R. A., Guizzardi, R. S. S. (2008), Grounding Software Domain Ontologies in the Unified Foundational Ontology (UFO): The case of the ODE Software Process Ontology , 11th Iberoamerican Conference of Software Engineering (CIbSE 2008), Recife, 2008. Guizzardi, R., Franch, X. and Guizzardi, G. (2012), “Applying a Foundational Ontology to Analyze Meansend Links in the i* Framework”, In: 6th IEEE International Conference on Research Challenges in Information Science, Valencia Spain, pp. 1 11. Guizzardi, R., Franch, X., Guizzardi, G. and Wieringa, R. (2013), “Ontological Distinctions between Meansend and Contribution Links in the i* Framework”, Lecture Notes in Computer Science, v. 8217, Heidelberg: Springer, pp. 463470. Kochanski, D. (2009), “Um Framework para Apoiar a Construção de Experimentos na Avaliação Empírica de Jogos Educacionais”, Master Dissertation in Applied Computing, UNIVALI, Brazil. Lucena, R., Santos, B., ; Silva, J., Silva, L., Alencar, R. and Castro, B. (2008), “Towards a Unified Metamodel for i*”. In: 2nd IEEE International Conference on Research Challenges in Information Science, Marrakech, v. 1, pp. 237246. Santos, B. (2008), “Istar Tool – Uma proposta de ferramentas para Modelagem de i*”, Master Dissertation in Computer Science, UFPE, Brazil. Travassos, G. (2002), “Relatório Técnico RTES590/02 – Introdução à Engenharia de Software Experimental”. Systems Engineering and Computer Science Program. COPPE/UFRJ, Brazil. Vokac, M. (2002), “Empiricism in Software Engineering: A Lost Cause?” Essay for MNVIT401. Yu, E. (1995), “Modelling Strategic Relationships for Business Process Reengineering”, Ph.D. thesis, Dept. of Computer Science, University of Toronto, Canada. Yu, E. (1997), “Towards Modelling and Reasoning Support for EarlyPhase Requirements Engineering”, In: 3rd IEEE International Symposium on Requirements Engineering. Annapolis, USA, pp. 226235.