=Paper= {{Paper |id=Vol-1835/paper02 |storemode=property |title=Quantitative Conceptual Model Analysis for Evaluating Simple Class Diagrams made by Novices |pdfUrl=https://ceur-ws.org/Vol-1835/paper02.pdf |volume=Vol-1835 |authors=Mizue Kayama,Shinpei Ogata,David K. Asano,Masami Hashimoto |dblpUrl=https://dblp.org/rec/conf/models/KayamaOAH16 }} ==Quantitative Conceptual Model Analysis for Evaluating Simple Class Diagrams made by Novices== https://ceur-ws.org/Vol-1835/paper02.pdf
Joint Proceedings of EduSymp 2016 and OSS4MDE 2016                                       Page 6




           Quantitative Conceptual Model Analysis
    for Evaluating Simple Class Diagrams made by Novices

 Mizue Kayama, Shinpei Ogata, David K. Asano & Masami Hashimoto
                                     Shinshu University

        {kayama, ogata, david, hasimoto}@cs.shinshu-u.ac.jp



       Abstract. In this paper, we aim to propose criteria for evaluating conceptual
       modeling errors made by university freshmen. We quantitatively analyzed class
       diagrams made by novice learners. Based on the results of three types of exper-
       iments, we propose 12 criteria, which are divided into 4 types, for evaluating
       class diagrams made by novices.

       Keywords: conceptual modeling, class diagram, criteria, quantitative analysis.


1      Introduction
These days, educational methods or learning courses related to conceptual modeling
have been explored in many educational institutes, academic conferences and aca-
demic journals [1-3]. The learners who were subjects of previous research were
mainly third or fourth-year undergraduate students, graduate students and/or young
engineers in computer science (CS). They already had finished many specialized
classes related to programming, object-oriented analysis and design, databases and so
on. Especially, almost all of the students were in a CS program as their major[4,5]. A
few studies whose subjects had little prerequisite knowledge, for example, high
school students (pre-university), undergraduate students in non-CS programs and CS
freshmen in their first semester, also reported their teaching experience or teaching
methods related to modeling education [6-8]. However, quantitative evaluation re-
sults were not shown in their reports. Especially for students in pre-university, Non-
CS and CS freshmen, there are no quantitative studies about model based thinking
and subsequent curricula of conceptual modeling.
     In this paper, we aim to propose criteria for evaluating conceptual modeling er-
rors made by university freshmen. To achieve this research goal, we quantitatively
analyzed class diagrams made by students. During this analysis, we asked ourselves
“What kind of criteria are suitable for novice learners when they create conceptual
models? ” and “Are there any differences between the scores of novice learners with
and without programming knowledge? ” There are two differences between previous
research and our research. The first difference is our subjects. We focus on universi-
ty freshmen and pre-university students, who have not taken any kind of CS specific
courses. The other difference is the empirical and continual research method. We
have been engaged in this research since 2010 [9,10].




                           Copyright © 2016 held by the author(s)
Joint Proceedings of EduSymp 2016 and OSS4MDE 2016                                 Page 7




2      Research Methods
2.1    Overview
We define conceptual modeling as a way of thinking to solve problems using engi-
neering methodology. The learning objectives of this course are to develop 3 types of
capabilities [11]:
1. Conceptual modeling: The ability to sketch a model diagram correctly according to
   a certain notation.
2. Requirements analysis: The ability to create the diagram so as to satisfy the re-
   quirements represented as sentences.
3. Appropriate abstraction: The ability to avoid defining unnecessary or inadequate
   classes and attributes for a target domain.
     Capability 1 is concerned with the ability to form concepts for designing visual
models. If learners lack this ability, they cannot read the given models correctly
based on requirements or cannot appropriately detect the differences between the
given model and the requirements. Capability 2 is related to the ability to capture the
essentials of software requirements. If learners lack this ability, they cannot create
suitable models for the requirements. Capability 3 is the same as the ability to ab-
stract fundamental features and/or significant entities from an object or service. If
learners lack this ability, they cannot control the abstraction level in model reading
and creation.

2.2    Simple Class Diagram
We use a class diagram which is a simplified standard class diagram defined using
UML2.x. Our simple class diagram has the minimum essential elements for concep-
tual modeling. For each class, a name and some attributes are listed, while no attrib-
ute types, method names, arguments, return types or visibilities are used. For each
association, two names and two multiplicities with four types (0..1, 1, 1..*, *) are used,
while no role, inheritance, aggregation, composition or dependency are used. The
only association used in this diagram is a simple association between two peer classes.
This association represents a pure structural relationship between two peers. Both
classes are conceptually at the same level, neither being more important than the other.
     In general modeling using object oriented methodology, classes in different lev-
els are used in one diagram. However, novice learners tend not to control abstraction
level appropriately. They often assign a system name to a class name or the name of
a concrete value to an attribute name. Therefore, we only used a subset of the nota-
tion of class diagrams from the original UML2.x notation.

2.3    Subjects
     Our subjects were 174 university students who were novices at conceptual mod-
eling. They were divided into two groups based on their computer science knowledge.
The members of the 11T group were 86 sophomores. They already had some com-




                          Copyright © 2016 held by the author(s)
Joint Proceedings of EduSymp 2016 and OSS4MDE 2016                                                                                        Page 8




puter science knowledge. Our experiment was held in their second semester in sec-
ond year during one of their elective courses. On the other hand, the subjects in the
12T group were 88 freshmen. They had not taken any CS related courses. Our exper-
iment for this group was held in their first semester in first year during one of their
required courses.
    All subjects were required to answer the questions individually. They were not
allowed to discuss the questions with each other or to solve the problems in groups.

2.4    Experimental Procedure
When humans acquire a new notation or concept, the typical first step is to read or
observe some appropriate samples. By doing this, some features of the notation or
concept can be captured. In the next step, the given notation or concept can be used
to draw or describe some product. Our experiment expands on this method by using
three tests: a model reading test, a model creation test and a model modification test.
     Before these tests, the instructor asked his students the names of the essential el-
ements in a simple class diagram to confirm their level of understanding. Two in-
structors were engaged in the course management. They planned the learning con-
tents of this course and gave our subjects lectures. Then, we analyzed the subjects’
answers and discussed the results.


3      Experimental Results
3.1    Model Reading Test
The goal of this test is to check the conceptual modeling capability of students. In
this test, students point out the differences between a given diagram and the problem
(P) statements. This test includes four problems. Each problem is related to classes,
attributes, associations and multiplicities. Among the choices, some statements are
not true for the given class diagram.
                              correct       incorrect                                         correct      incorrect

      P4                 69.8%                          30.2%           P4                 72.7%                          27.3%


      P3                  74.4%                          25.6%          P3                    86.4%                          13.6%


      P2        32.6%                         67.4%                     P2         46.6%                          53.4%


      P1                         91.9%                          8.1%    P1                     92.0%                              8.0%

           0%     20%        40%         60%        80%          100%        0%   20%        40%         60%        80%            100%
                    % of correct and incorrect answers                              % of correct and incorrect answers

                            (a) 11T group                                                  (b) 12T group


      Fig. 1. % of problems answered correctly and incorrectly in the model reading test

     Figure1 shows the percentage of problems answered correctly and incorrectly in
this test for the two groups. The trend of the percentage of problems answered cor-
rectly is the same for both groups. The students’ level of understanding decreases as
follows: class > association > multiplicity > attribute. The average total scores and in
particular the attribute related problem scores of these two groups show a significant




                                         Copyright © 2016 held by the author(s)
Joint Proceedings of EduSymp 2016 and OSS4MDE 2016                                        Page 9




difference. So, the level of understanding about “attribute” is much lower than for
other elements (class, association, multiplicity) for both groups.

3.2     Model Creation Test
The goal of this test is to check the requirement analysis capability and the appropri-
ate abstraction capability. The 12T group (38.6%) has a higher score than the 11T
group (14.0%). However, the average total scores and variances of these two groups
show a significant difference. At first, we categorized the answers which had some
errors based on the three error types from previous research. As a result, we found a
new error type: class related error. In total, we extracted four types of errors: syntac-
tic errors, attribute related errors, association related errors and class related errors.
Figure 2 shows the percentage of the four error types that occurred in the model crea-
tion test. For the 11T students, the number of incorrect answers was 74. For the 12T
students, the number of incorrect answers was 54.
                     100%
                      90%                  97.3% 96.3%
                                                                        11T      12T
                      80%
                      70%
                      60%
                      50%
                      40%
                      30%                                       33.3%
                      20%
                      10%                                6.8%           25.7%
                            5.4%   11.1%                                        5.6%
                       0%
                             Syntactic       Attribute   Association        Class
                              errors          related     related          related
                                               errors      errors           errors

            Fig. 2. % of the four error types that occurred in the model creation test.
     In this experiment, we found that attribute related errors are the most common
type of error made. In both groups, over 95% of the incorrect answers had this type
of error. For the 11T group, which has programming knowledge, the percentage of
class related errors is relatively higher. On the other hand, 12T group, which has no
programming knowledge, shows a higher percentage of association related errors.
     We analyzed these three in four types of errors in more detail.
     The class related error has two detailed subcategories:
   (a) There are some classes which have different abstraction levels in one diagram.
   (b) There are more than two classes whose names or attributes have the same
       meaning in one diagram.
      The attribute related errors had six detailed error categories:
   (a) A class does not have any attributes (No attribute). This error also is included
       the syntactic error.
   (b) Two or more classes have the same set of attributes (Same attribute). “Same”
       means that each attribute has the same range of values.
   (c) An attribute is defined as not “name” but “value” (Value attribute).
   (d) Attributes which are actions or methods are listed (Behavioral attribute).
   (e) The meaning of both an attribute and the multiplicity of an association is over-
       lapped (Overlapped property).




                             Copyright © 2016 held by the author(s)
Joint Proceedings of EduSymp 2016 and OSS4MDE 2016                                                             Page 10




   (f) Duplicated attributes are used (Duplicated attribute).
    The association related error type includes some class diagrams which have no
association name or multiplicity and have inadequate association name or multiplicity.
This type has four detailed error categories.
   (a) There are no association names. This error also is included the syntactic error.
   (b) Inadequate association name is given.
   (c) There are not two multiplicities for one association. This error also is included
       the syntactic error.
   (d) Inadequate multiplicity is given.
     Figure 3 shows the percentage of attribute related error types in the model crea-
tion test. For the 11T students, the number of students who made attribute related
errors was 72. For the 12T students, the number of students who made attribute relat-
ed errors was 52.
                 100%
                 90%
                                                                                                 11T   12T
                 80%
                 70%
                 60%                                   56.9%

                 50%
                 40%
                 30%                   26.4%                   59.6%
                        19.4%                                                       20.8%
                 20%
                                               32.7%
                 10%
                                9.6%                                   1.4% 1.9%                       1.9%
                  0%
                            No            Same           Value         Overlapped   Behavioral    Duplicated
                         attribute      attribute       attribute       property     attribute     attribute


              Fig. 3. % of attribute related error type in the model creation test

     In the 6 subcategories of the attribute related error, the “Value attribute” errors
and the “Same attribute” errors have relatively high occurrence percentages for both
groups. For the 11T, which has programming knowledge, the percentage of the “No
attribute” errors and the “Behavioral attribute” errors is about 20%. The “Duplicated
attribute” errors occurred only in the 12T, which has no programming knowledge.

3.3    Model Modification Test
Overview. The goal of this test is to check the ability of conceptual modeling, re-
quirement analysis and appropriate abstraction. In all five problems, students need to
point out the mistakes in each class diagram and describe why they are incorrect.
Then, they are asked to modify the class diagram to correct the mistakes. P1 has
association related errors, which are inadequate multiplicity and duplicate association
names. P2 has an attribute related error, where the attribute name is defined as a
value instead of a property. P3 has an association related error, which is inadequate
multiplicity. P4 has an association related error, whi ch is the lack of association
names. P5 has a syntactic error, which is redundant multiplicity.

Results. Figure 4 shows the percentage of questions answered correctly and incor-
rectly in the model modification test for the two groups. The trend of the percentage
of questions answered correctly is the same for both groups. The highest percentage




                                 Copyright © 2016 held by the author(s)
Joint Proceedings of EduSymp 2016 and OSS4MDE 2016                                                                          Page 11




of correctly corrected errors was for the syntactic error (P1) and the association relat-
ed error (P4). The lowest percentage was for the association related error (P3). Their
level of understanding decreases as follows: P1 > P4 > P2 > P5 > P3. Both P1 and P4
are lacking necessary elements in the diagram. P2, P5 and P3 have inadequate ele-
ments in the given diagrams. This means that the “inadequate description” error is
more difficult to modify than the “lack of necessary element” error. The average total
scores and variances of these two groups are statistically the same. Only the P3
scores of these two groups show a significant difference.
                              correct    incorrect                                          correct        incorrect

    P5                        80.2%                          19.8%      P5                   84.0%                        16.0%

    P4                           91.9%                           8.1%   P4                     91.5%                        8.5%

    P3                55.8%                          44.2%              P3                 76.6%                       23.4%

    P2                        82.6%                          17.4%      P2                    88.3%                        11.7%

    P1                           91.9%                           8.1%   P1                      94.7%                          5.3%

         0%      20%        40%         60%        80%           100%        0%   20%        40%         60%        80%        100%
                   % of correct and incorrect answers                               % of correct and incorrect answers
                              (a) 11T group                                                (b) 12T group

      Fig. 4. % of questions answered correctly and incorrectly in the model modification test.


4             Discussion

4.1           Question 1: What kinds of criteria are suitable for novice learners when
              they create conceptual models with simple class diagrams?
     We propose 12 criteria, which are divided into 4 types, for evaluating simple
class diagrams made by novices for conceptual modeling based on the results we
mentioned above. Table 1 shows the proposed criteria. The frequency of occurrence
is different for each item. However, by using these items we can check the level of
understanding for conceptual modeling of novice learners. Therefore, conceptual
modeling instructors can develop their course for novices with these criteria.
Table 1. Criteria for evaluating simple class diagram made by vovice for conceptual modeling

    Error types                                                 Criteria
                         Inadequate notation.
         Syntactic
                         Lack of necessary element [association and multiplicity].
           error
                         Lack of name of necessary element [class, attribute and association].
                         There are some classes which have different abstraction levels in one diagram.
    Class related
       error             There are two or more classes whose names or attributes have the same meaning in
                         one diagram.
                         The same attributes are used in more than two classes.
                         The attribute name is defined as a value instead of a property.
        Attribute
                         An attribute that describes an action is used .
      related error
                         An attribute that has the meaning of multiplicity is used.
                         Duplicated attributes are used.
     Association         Inadequate association name.
    related error        Inadequate multiplicity.




                                              Copyright © 2016 held by the author(s)
Joint Proceedings of EduSymp 2016 and OSS4MDE 2016                               Page 12




4.2    Question 2: Are there any differences between the programming-known
       group and the not-known group in terms of their level of understanding of
       conceptual modeling?
     In the model reading test, there are no significant differences between the 11T
group and the 12T group in terms of the percentage of questions answered correctly.
Especially, the trend of the percentage of problems answered correctly is statistically
the same for both groups. The average of the percentage of questions answered cor-
rectly by the 12T group is higher than the average of the 11T group. However, there
are no statistically significant differences between any scores.
     In the model creation test, there is a statistically significant difference between
scores of the 11T group and the 12T group.
   (a) The percentage of problems answered correctly by the 12T group which has no
       programming knowledge is higher than the 11T group which has programming
       knowledge.
   (b) About the percentage of the four error types that occurred in this test, for the
       11T group the percentage of class related errors is high. On the other hand, the
       12T group shows a high percentage of association related errors.
   (c) About the percentage of attribute related error types in this test, for the 11T
       group, the percentage of no attribute errors and behavioral attribute errors is
       about 20%. Duplicated attribute errors occurred only in the 12T group.
     About (a), though these two groups were given the same contents and the same
length of lectures about conceptual modeling, the 12T group which has no program-
ming knowledge showed a higher score in creating models based on the given re-
quirements. About (b), whereas the association related errors are relatively superficial
mistakes, the class related errors are quite essential mistakes in conceptual modeling
using class diagrams. These types of errors are concerned with abstraction level con-
trol. This fact means that programming knowledge has no effect on the ability to
control abstraction levels. About (c), the behavioral attribute errors occurred only in
the 11T group. We think this fact is caused by structured programming knowledge
which includes functions. If they draw a class diagram with methods, students in the
11T group would get a higher score on this test. Therefore, it is better to teach con-
ceptual modeling with this notation before programming. However, the total trend of
our 13 criteria seems to be the same for both groups.
     The results of the model modification test are the same as the model reading test.
There are no significant differences between the 11T group and the 12T group in
terms of the percentage of questions answered correctly. Especially, the trend of the
percentage of problems answered correctly is statistically the same for both groups.
The average of the percentage of questions answered correctly by the 12T group is
higher than the average of the 11T group. However there are no statistically signifi-
cant differences between any scores. Overall, based on our experiments, program-
ming knowledge seems to not directly affect conceptual modeling ability. If so, con-
ceptual modeling education in this notation for university freshmen is reasonable. In
this case, the instructors should consider our 12 criteria listed above.




                          Copyright © 2016 held by the author(s)
Joint Proceedings of EduSymp 2016 and OSS4MDE 2016                               Page 13




5      Conclusion
Our research questions are “What kind of criteria are suitable for novice learners
when they create conceptual models?” and “Are there any differences between the
scores of novice learners with and without programming knowledge?”
     In this paper, we propose criteria for evaluating conceptual modeling errors made
by novices based on the results of three experiments. We found that there is no rela-
tion between programming knowledge and conceptual modeling ability for the nota-
tion used in our experiments. We used real world objects in our models, not abstract
objects in this study. Moreover, we asked students to solve each problem individually,
without discussion with other students.
     The effects of these matters for the proposed conclusions need to be considered
in future work. Also, we need to discuss the relation between diagram notation and
education timing more carefully.


       Acknowledgement
This work was supported by JSPS KAKENHI Grant Number 22300286 & 16H03074.


       References
 1. J.Bezivin, et al, “Teaching Modeling: Why, When, What?”, MODELS 2009, 55-62 (2009)
 2. J.Börstler, I.Michiels, and A.Fjuk, “ECOOP 2004 Workshop Report: Eighth Workshop on
    Pedagogies and Tools for the Teaching and Learning Object-Oriented Concepts”, ECOOP
    2004 Workshop Reader, 36-48 (2004)
 3. J.Kramer,J. “Is Abstraction The Key To Computing?”, CACM, 50(4), 37-42 (2007)
 4. N.Bolloju and F.S.K.Leung, “Assisting Novice Analysts in Developing Quality
    Conceptual Models with UML”, CACM, 49(7), 108-112 (2006)
 5. H.C.Cham, H.H.Teo and X.H.Zeng, “An evaluation of Novice End-User Computing
    Performance : Data Modeling, Query Writing and Comprehension”, J. of the American
    Society for Information Science and Technology, 56(8), 843-853 (2005)
 6. J.Niere and C.Schulte, “Avoiding anecdotal evidence : An experience report about
    evaluating an object-oriented modeling course”, MODELS 2005 Educator's Symposium,
    63-70 (2005)
 7. C.Schulte and J.Niere, “Thinking in Object Structures: Teaching Modeling in Secondary
    Schools”, 6th ECOOP Workshop on PTLOOC (2002)
 8. S.Akayama et al, “Development of a Modeling Education Program for Novices using
    Model-Driven Development”, Workshop on Embedded and Cyber-Physical Systems
    Education (2012).
 9. Y.Hara. et al, “A Basic Study of an Educational Environment for Modeling and
    Abstraction”, ESS 2011, pp.16-1-16-5, 2011.
10. M.kayama et al, “A Practical Conceptual Modeling Teaching Method Based on
    Quantitative Error Analyses for Novices Learning to Create Error-Free Simple Class
    Diagrams”, 3rd ICAAI, 616 - 622 (2014)
11. S.J.Merror and M.C.Balcer, “Executable UML: A Foundation for Model Driven
    Architecture”, Addison-Wesley Professional (2002)




                          Copyright © 2016 held by the author(s)