=Paper= {{Paper |id=Vol-490/paper-4 |storemode=property |title=A Measurement-Based for Adopting Usability Engineering Methods and Tools |pdfUrl=https://ceur-ws.org/Vol-490/paper_04.pdf |volume=Vol-490 |dblpUrl=https://dblp.org/rec/conf/iused/MetzkerS09 }} ==A Measurement-Based for Adopting Usability Engineering Methods and Tools== https://ceur-ws.org/Vol-490/paper_04.pdf
                     Adoption of Usability Engineering Methods:
                          A Measurement-Based Strategy
                       Eduard Metzker                                                  Ahmed Seffah
            Vector Informatik, Stuggart Germany                              EHL-LHR, Lausanne Switzerland
                 eduard.metzker@gmx.net                                          seffah.ahmed@ehl.ch


ABSTRACT                                                           Furthermore, there is little hard evidence backing up new
In the context of a software development organization, two         technology to be adopted, and their costs and benefits are
strategies are possible for introducing and institutionalizing     rarely understood [1]. Without this data, choosing a particular
new usability engineering methods. The first one, expert-based     technology or methodology for a project at hand essentially is
institutionalization, require to resort to third party companies   a random act with many consequences [2]. The findings from
or experts that can, based its previous expertise, assist the      a very large survey made by Standish group, new technology
team in selecting, implementing and institutionalizing             is one of the top ten reasons for projects failure or success
usability engineering methods and tools. The second one, a         [27].
measurement-based strategy, is based on empirical evidence
for learning and assessing the appropriateness, usefulness of a    In order to support and effective adoption, a new metrics-
usability engineering method. This paper proposed to combine       based approach comprising a process model and a support
these two approaches in a single process metrics support           environment are presented in this paper. A large case study
environment for selecting and institutionalizing usability         was developed to assess the acceptance of the approach by
engineering methods. The proposed approach has been                development teams. The evaluation involved 44 professional
validated via in a cross-organizational empirical study            software engineers from five medium to large-sized
involving several software engineers from five mediums to          organizations. The evaluation method is discussed including
large sized software development companies.                        context, method, subjects, procedure and results. Implications
                                                                   of the results on the design of metrics-based strategy are
                                                                   discussed for adopting new technology and assessing their
Keywords                                                           acceptance by project teams. Based on the results, a set of
Metrics,     usability,   usability    engineering   methods,      guidelines is derived to optimize the acceptance of metrics
institutionalization,   adoption,     software   developments      exploitation approaches by project personnel.
organization
                                                                   2. THE PROPOSED                  METRICS          SUPPORT
1. INTRODUCTION
                                                                   ENVIRONMENT
Within the scope of this research, by adoption we refer to the
process and the related tools for selecting the appropriate new    The overall goal of the proposed approach, called Adoption-
software development technology while assessing their              Centric Usability Engineering (ACUE), is to facilitate the
suitability to the project needs and size as well as the           adoption of UE methods by software engineering practitioners
capability of the personnel to use effectively and efficiently     and thereby improve their integration into existing software
the new established technology. Adoption has been always a         development methodologies and practices. ACUE is designed
key challenge for software development organizations [28]. It      to support project teams in institutionalizing this abstract
was reported that the management staff commitment and the          knowledge about UE methods and to help them transfer this
involvement of the personnel represent the top factors that        knowledge into their development processes. UE methods are
impact on the success with a new technology when first             perceived as integrated into an existing software development
introduced [27, 29, and 30].                                       process when they are adopted by the project team, i.e. when
                                                                   they are accepted and performed by the project team.
However, despite management efforts made by organizations
to render the transition more “user-friendly”, the associated      ACUE exploits empirical data collected in different projects to
help and training material, although precise and perfectly         yield stronger evidence on how the method works in a certain
describing the new method are often delivered in an esoteric       context. The data form an empirical base to guide the
and unreadable language. Another important factor is that          improvement of UE methods and to facilitate the informed
organizations and managers are usually overly optimist in          selection and deployment of UE methods in future projects. If
their employees’ ability to quickly master a new technology.       this approach is applied repeatedly in a number of projects
The reality is that understanding how to apply the technology      over time, it leads to an incremental construction of a body of
is a long and arduous process.                                     evidence to guide usability engineering method selection
                                                                   (Figure 1).
                                                                   increase its probability of selection in similar projects in the
                                                                   future.

                                                                   The characteristics of the project are specified using the
                                                                   context models stored in the repository. A context model
                                                                   includes context factors to describe various project constraints
                                                                   such as the resources available in the project or the type of
                                                                   product to be developed. Each context model consists of a set
                                                                   of factors that can have nominal, ordinal or interval scale
                                                                   measures [11]. An example for an ordinal factor that describes
                                                                   a property of a product to be developed is ‘user interface
                                                                   interaction complexity’. This factor may have three
                                                                   characteristics ‘text-based interface’, ‘graphical user interface’
                                                                   or ’multi-modal interface’. Depending on the project
                                                                   characteristics, appropriate methods are suggested by the
                                                                   system. Candidate methods are ranked according to two
                                                                   different criteria: (1) similarity between the method and the
                                                                   project characteristics, (2) results of assessments from project
                                                                   teams that used the methods in previous projects.

                                                                   Within, the system the usability engineering methods are
                                                                   provided in the format of method packages. Each package
                                                                   contains a textual description of a method that is structured
                                                                   according to the pyramid principle [12]. Auxiliary material
                                                                   such as templates, checklists and tools is linked to each
Figure 1: The overall view of ACUE approach                        package. This material facilitates easy compliance of the
                                                                   method described. The process guidance remains passive and
The support environment is called ESPrEE (Evidence-based           does not enforce the performance of the methods proposed.
Software PRocess Evolution Environment). The components
of ESPrEE are integrated via a web portal and they be              The constraints of industrial software development projects
remotely accessed using any web browser. Its core                  often enforce the invention of new methods or the adaptation
functionalities are:                                               and streamlining of existing methods. For these reasons the
                                                                   environments provides means for capturing methods and
•    At the beginning of a new project, the metrics-based          integrating them into the repository.
     method selection component of the environment is used
     to configure the set of usability engineering methods that    3. EVALUATION
     will be applied in the project
•    After the method selection has been completed, the            3.1. Specific purpose of the evaluation
     environment generates a project-specific hyper-media
     workspace in which the methods selected are graphically       A large number of measurement programs are suspended or -
     visualized according to the project phases in which their     in the worst case – failed. This is because the measurement
     usage is intended                                             program is not accepted by stakeholders for the following
•    At the completion of major project phases or in post          reasons [3, 4, 5, 6 and 7]:
     mortem sessions [13], the quality of the methods
     employed is assessed by the project team against a            1.   The measurement process is perceived as tedious and
     quality model. For this purpose quality models contain a           time consuming
     set of quality factors and carefully defined rating scales    2.   Effort and benefits of the program are poorly distributed
                                                                   3.   The impact on daily practice is perceived as being too
The core of the system is a fuzzy multi-criteria decision-              low to justify sustained effort
making engine. The characteristics of the usability methods        4.   Metrics support tools and/or processes are difficult to use
and projects are defined as sets of fuzzy sets. Based on these
models the engine is able to compute similarity measures for       To examine if the proposed approach addresses these issues,
projects and methods to facilitate decisions based on analogy.     the evaluation study was designed with the following
The engine is coupled with and assessment component. If a          questions in mind:
method receives a poor assessment in a certain project context,
the method’s characteristics are automatically adapted to          •    Does each member of the project team understand the
reduce the probability of the method being selected in similar          basic principles and structure of the method without
projects. On the other hand, if a method has successfully been          extensive training?
applied in a certain project, its characteristics are adapted to
•    How do project managers assess the potential quantitative       3.4 Method
     effects of the approach on their practices? Would they          In this study, Davis’ technology acceptance model (TAM)
     use the approach in their setting?                              [14] was used. TAM’s assessment dimensions ‘perceived
•    Which problems the project personnel may face when              utility’ and ‘perceived ease of use’ were extended while
     applying the metrics support tool underlying the proposed       adding ‘understandability’ as a third dimension. TAM
     framework?                                                      postulates that tool acceptance can be predicted by measuring
                                                                     two dimensions: the perceived usefulness (PU) and the
3.2. Context of the evaluation                                       perceived ease-of-use (PEU) of a system.

We used a set of five medium- to large-size software                 The perceived usefulness of the system expresses the
engineering companies developing advanced next-generation            “subjective probability that using a specific application
home entertainment systems, driver assistance technology for
                                                                     system will increase (the user’s) job performance within
passenger cars and military support systems. The usability of
these systems has been recognized by the organizations as a          an organizational context”, i.e. it is a measure for the
crucial quality factor. While usability engineering methods          perceived utility of the system.
[10] are well-know by these companies to ensure the
development of software with high usability, no experience           The perceived ease-of-use is the main factor that
with usability engineering was available in the engineering          influences the acceptance of a system. Davis defines
teams. ESPrEE was configured for this environment.                   perceived ease-of-use as “the degree to which the user
Appropriate context and quality models were defined and              expects the target system to be free of effort”, i.e. it is a
usability engineering methods were captured in method                measure for the usability of a system.
packages. Resources included successful methods invented in
previous industrial engineering projects such as reported in
[16], methods distilled from literature on usability engineering     Together, perceived ease-of-use and perceived
[10, 17, 18], and recent research results such as Spencer’s          usefulness constitute the person’s attitude toward using
‘streamlined cognitive walkthrough’[19]. This initial                a system. The attitude (A) and the perceived ease-of-use
population of the support tool was performed by a team of            (PEU) influence the behavioral intention (BI) which can
professional usability engineering experts and took about 2.5        be used to predict the actual use of the system. The
man-months of effort. The participating organizations were           technology acceptance model (TAM)
part of a government-supported software engineering research
consortium. However, no organization committed to adopt the          TAM’s dimension of perceived utility was further divided into
approach prior to the evaluation.                                    the following sub-dimensions:

3.3 The Subjects                                                     •    Perceived compatibility with daily practice
                                                                     •    Perceived increase in efficiency
All 44 subjects participated in the study on a voluntary basis.      •    Perceived usefulness
Of them, 39 are full-time employees working as software              •    Perceived support of working tasks
engineers for the companies described in section 3.2. Five
subjects were graduate students working for the companies on         The sub-dimension of perceived utility was measured by
a part-time basis. The subjects were involved in developing          qualitative effects analysis while usability was examined by
highly interactive software systems in a variety of domains,         studying user behavior during the user’s interaction with the
e.g. driver assistance systems, home entertainment, and              metrics support environment [10]. Understandability was
military defense systems. Based on their experience with the         examined via a knowledge test in which the subjects answered
development of highly interactive systems, the subjects were         questions on the concepts of the overall approach and the
classified into three groups: new employees (NE, 10),                metrics support environment. The knowledge test was
software engineers (SE, 21), and usability engineers (UE, 13)        performed before and after the interaction with the tool to
[10]. In the following, these groups are referred to as user         study if and how the understanding of the concepts by the
groups for reasons of simplicity.                                    subjects changes.

                                                                     To implement the study two basic techniques were deployed:
                                                                     questionnaires and scenario-based task solution. The purpose
                                                                     of the two questionnaires deployed (q1 and q2) are
                                                                     summarized in table 1.
Table 1: Purpose of the questionnaires deployed
                Data collected
      q1        •    Subject characteristics (age, qualification, professional experience)
                •    Typical role of subject in the software engineering process
                •    The subjects knowledge on the concepts of the approach and the metrics environment (pre-test, based
                     on the introductory material supplied)
      q2          •    The subjects knowledge on the concepts of the approach and the metrics environment (post-test,
                       based on the scenario-guided interaction with the metrics environment)
                  •    The subjects assessment of the utility and usability of the metrics environment

The scenario-based tasks that were specific for each user              4.5 Tasks
group forms the core of the evaluation. While the subjects             To identify potential problems and study the acceptance of the
were working on a task, behavioral data and any comments               metrics support environment under realistic conditions,
made while thinking out loud were captured as a basis for              specific usage scenarios were developed for each user group.
improving the metrics environment. Section 4.5 describes the           Each scenario reflects the role of the subject and details the
tasks that were performed by the subjects while the exact              tasks to be solved.
evaluation procedure and deployment of the methods is
described in section 4.6.                                              All scenarios were embedded in a cover story that set a
                                                                       common background for all scenarios. Scenario S0 is
                                                                       equivalent for all user groups. In S0, the subjects were allowed
                                                                       to freely explore all components of the metrics environment.
                                                                       The other tasks to be solved in each scenario are different
                                                                       among the user groups (Table 2).

Table 2: Tasks to be solved in scenarios
                      Tasks
    NE-S1             Find an explanation of the term usability engineering. Mark the section for later exploration.
    NE-S2             Find an introductory article about evaluation and tests. Review the material supplied.
    NE-S3             Open the hypermedia workspace for the project DIGital. Find and open the method package on
                      heuristic evaluation.
    SE-S1             Browse all method packages available for project DIGital. Find and display a package on heuristic
                      evaluation. Assess the quality of the method heuristic evaluation.
    SE-S2             Comment the method ‚heuristic evaluation‘. Edit the method ‚heuristic evaluation‘. Extend the
                      method package with a checklist to be used in ‚heuristic evaluation‘.
    SE-S3             Create a new method package. Fill the package with given raw input material. Specify the meta-
                      data of the methods context model. Link the package to related packages.
    UE-S1             Create a new project PORTAL. Choose a context model and specify the project characteristics via
                      the context model. Choose appropriate method packages based on the project characteristics
                      specified. Trigger generation of hypermedia workspace for the project PORTAL.
    UE-S2             Manage access to the hypermedia workspace for the project PORTAL. Invite project team
                      members. Manage access levels of project team members.

3.6. Test procedure and scripts                                        the tasks. After the tasks of all scenarios were completed,
                                                                       questionnaire q2 was handed out to the subject. The evaluation
Prior to the evaluation sessions, all the subjects received            is concluded with a brief free discussion.
introductory material. It described the concepts of the metrics
approach and the components of the related environment.                Two observers were involved in each evaluation session. One
Single subjects, who, for some reason, had no access to the            observer recorded the behavioral data, while the other was
material prior to the evaluation, were given the opportunity to        responsible for writing down comments from the subjects
study printouts of the material. Each subject had an individual        were thinking aloud. During the session, the subjects worked
session, no group sessions were performed.                             with a laptop with each evaluation lasting of roughly two
                                                                       hours.
The evaluation session started with a short introduction of the
participants, the procedure, and the objectives of the study.          4. RESULTS AND FINDINGS
The subjects were explicitly informed that the goal of the
evaluation was to assess the utility of the approach and not the       The qualitative effects analysis shows that the proposed
capabilities of the participants and that all data would be            metrics-based strategy is strongly accepted by the main target
treated confidentially. First questionnaire q1 was handed out.         group. However the support environment receives higher-
Next the tasks to be solved were handed out in form of                 than-average ratings from all subject groups.
scenarios. Scenario S0 was performed by each participant to
promote a free exploration of the system. The time for S0 was          4.1 Understandability of the approach
limited to 20 minutes. Next, the group-specific scenarios were
handed out to the subjects. No time limit was set for task             The understanding of the metrics collection approach and the
completion. The subjects were encouraged to articulate                 metrics support environment by the subjects was measured
impressions and problems and think aloud while performing              before and after usage of the support system via the
questionnaires q1 and q2. After reading the introductory
material, the average percentage of correct answers was 36%.                                                3,4
                                                                      NE                                                   4,3
                                                                                                            3,4
Subsequent to performing the scenario-based tasks, this value                                                3,5
                                                                                                               3,6
doubled, being up to 63%. The performance of the groups and
the overall performance of the subjects are depicted in figure                                          3,3
                                                                      SE                                                    4,4
2. It shows that even the relatively short time of usage of the                                       3,1
                                                                                                             3,5
metrics environment led to a significant increase in the                                                    3,4
understanding of the concepts and components of the                                                               3,7
approach. The increased understanding of the overall approach         UE                                                      4,5
                                                                                                                     3,9
                                                                                                                        4,1
was lowest in the group of new employees (NE). However this                                                                4,3
can be easily explained since their scenarios (NE-S1-S3) did
                                                                                                            3,4
not comprise the usage of all components of the metrics               ∅                                                     4,4
                                                                                                        3,3
support system.                                                                                                   3,7
                                                                                                                  3,7

                                                                           1          2           3                  4              5
                                  43%
  NE                                48%                                             compatibility with daily practice
                            35%                                                     increase in knowledge
  SE                                               71%                              usefulness as a tool
                            35%                                                     increase in the efficiency of UE
  UE                                         62%
                                                                                    support of task

  ∅
                            36%                                    Figure 3: Results of the qualitative effects analysis
                                             63%                   (assessment scores for utility dimensions with respect to
                                                                   user group, 1: very low, 5: very high)
      0%       20%          40%        60%          80%

                post-test         pre-test                         The results indicate that the approach and its support
                                                                   environment were generally assessed by all groups as higher-
                                                                   than-average useful and reasonable. All subjects seem to
Figure 2: Pre- and post-test scores in knowledge tests with        highly appreciate the potential of the approach for increasing
respect to subject groups (in percentage of correctly              their knowledge of usability engineering methods. The
answered questions)                                                dimension ‘task support’ receives the highest scores from the
                                                                   UE group. This corresponds with the pre-configuration of the
4.2 Usefulness of the approach                                     support environment with usability engineering methods and
                                                                   the subjects role as usability engineers. It could be concluded
For the qualitative effects analysis, the subjects were asked to   that the assessment of this dimension by the other groups
assess the properties of the metrics support environments          could be further enhanced, if also usability engineering
along with the utility dimensions defined in section 3.4Each       methods for other areas such as ‘requirements engineering’ or
subject filled out questionnaire q2 after performing the           methods for enhancing programmer productivity were
scenario-specific tasks. For this reason questionnaire q2          integrated into the method repository of the metrics
includes a number of items to be rated on a five-level Likert      environment. This result underpins the necessity to provide
scale [20] for each dimension. Figure 3 sets out the results of    benefits for all groups of project personnel involved in metrics
the qualitative effects analysis. The bars represent ratings of    collection and exploitation.
the assessment dimensions. The mean ratings were calculated
for each dimension and grouped according to the subject            4.3   Recommendations                             for          usability
groups.
                                                                   improvements
                                                                   The behavioral data and user comments recorded during task
                                                                   performance suggest that there is potential for improving the
                                                                   usability of the metrics support environment. The distribution
                                                                   of usability issues identified by subjects across the client
                                                                   components of the metrics support environment are presented
                                                                   in figure 4.

                                                                   Most improvement suggestions are related to the components
                                                                   for process guidance and metrics-based decision support. The
                                                                   high number of usability issues identified for the process
                                                                   guidance component can be partially explained by the fact,
                                                                   that the scenarios of all user groups (NE, SE, UE) included
                                                                   interaction with the process guidance component.
                                                                    deploying the methods in the project. Finally the project
             7%                                                     assesses the quality of the methods deployed. The approach is
                                                                    supported by a tool, an intranet that offers a web-based
                  11%
                                                                    support system.       The approach has been successfully
                            20%                                     implemented in industry.
                                           32%
                                                                    Instead of evaluating the approach in an isolated longitudinal
                                       30%                          case study, a study was performed to examine the acceptance
                                                                    of the approach by practitioners from various organizations.
  0%         10%         20%         30%         40%                The acceptance was measured in scenario-driven evaluation
       cross component issues                                       sessions, by capturing the understandability, perceived utility
       web portal                                                   and perceived usability of the approach. The results indicate
       method assessment and improvement component
       process guidance component                                   that the approach is accepted by the subject groups examined.
       metrics-based method selection component                     We recommend using the study described as a template to
                                                                    estimate the initial acceptance when introducing tool
Figure 4: Distribution of the usability issues identified over      supported measurement programs into organizations. Such
client components of metrics support environment                    studies can be useful for early identification of acceptance
                                                                    problems that hamper the log-term success of metrics
One issue is that more assistance in working with the UE            programs. The usability of the metrics environment will be
methods is appreciated by the subjects. In particular novice        further improved using the feedback gathered.
users could profit from concepts such as wizards to augment
metrics capturing. Moreover the consistency between the             6. ACKNOWLEDGEMENTS
applications should be increased. Finally the parts of the
terminology used in the graphical user interfaces of the            Part of the empirical study presented in this paper was
metrics environment should be revised for more                      originally conducted at Daimler Chrysler. We would like to
comprehensible terms. One example given was that subjects           thank Elke Wetzstein and Gerd Jan Tschoepe from the
suggested changing the term “project context model” to              Institute for Industrial Psychology of the Humboldt University
“project characteristics”.                                          Berlin for their support in preparing and conducting the
                                                                    evaluation sessions. Part of this research work was funded by
                                                                    the BMBF (German Ministry for Research and Education)
                                                                    under the project EMBASSI (01IL904I). We would like to
5. A CONCLUDING REMARK                                              thank also the National Science and Engineering Research
                                                                    Council of Canada and Daimler Chrysler for their financial
In this paper, an approach for adopting UE methods was              support to the human-centered software engineering group at
proposed. It consists of a three-phased process. First, usability   Concordia University.
engineering methods are selected for a new project based on
the projects constraints. Then the project team is supported in

                                                                    [5]      R. T. Hughes, “Expert Judgment as an Estimating
REFERENCES                                                                   Method,” Information and Software Technology, pp.
                                                                             67-75, 1996.
[1]       D. E. Perry, A. A. Porter, and L. G. Votta,               [6]      F. Niessink and H. Van Vliet, “Measurements
          “Empirical Studies of Software Engineering: A                      Should Generate Value, Rather than Data.,” in Proc.
          Roadmap,” in Proc. International Conference on                     Sixth IEEE International Symposium on Software
          Software Engineering (ICSE2000), 2000, pp. 345 -                   Metrics, 1999, pp. 31-39.
          355.                                                      [7]      O. Laitenberger and H. M. Dreyer, “Evaluating the
[2]       V. R. Basili, F. Shull, and F. Lanubile, “Building                 Usefulness and the Ease of Use of a Web-based
          Knowledge Through Families of Experiments,”                        Inspection Data Collection Tool,” in Proc. Fifth
          IEEE Transactions on Software Engineering, vol.                    IEEE International Symposium on Software Metrics,
          25, pp. 456-473, 1999.                                             1998.
[3]       D. R. Goldenson, A. Gopal, and T. Mukhopadhyay,           [8]      S. Komi-Sirviö, P. Parviainen, and J. Ronkainen,
          “Determinants of Success in Software Measurement                   “Measurement        Automation:     Methodological
          Programs: Initial Results,” in Proc. Sixth IEEE                    Background and Practical Solutions - A Multiple
          International Symposium on Software Metrics,                       Case Study,” in Proc. 7th IEEE International
          1999, pp. 10-21.                                                   Software Metrics Symposium, 2001, pp. 306-316.
[4]       B. G. Silverman, “Software Cost and Productivity          [9]      L. Rosenberg and L. Hyatt, “Developing a
          Improvements: An Analogical View,” IEEE                            Successful Metrics Program,” in Proc. 8th Annual
          Computer, vol. May 1985, pp. 86-96, 1985.                          Software Technology Conference, 1996.
                                                                    [10]     J. Nielsen, Usability Engineering: Morgan Kaufman
                                                                             Publishers, 1994.
[11]   L. Briand, K. El Emam, and S. Morasca, “On the         [20]   C. M. Judd, E. R. Smith, and L. H. Kidder, Research
       Application of Measurement Theory in Software                 Methods in Social Relations, 6 ed: Harcourt Brace
       Engineering,” Empirical Software Engineering, vol.            Jovanovich College Publishers, 1991.
       1, pp. 61-88, 1996.                                    [21]   G. F. Smith and S. T. March, “Design and Natural
[12]   B. Minto, The Pyramid Principle - Logic in Writing            Science Research on Information Technology,”
       and Thinking, 3rd ed. London: Minto International             Decision Support Systems, vol. 15, pp. 251-266,
       Inc., 1987.                                                   1995.
[13]   A. Birk, T. Dingsøyr, and T. Stålhane, “Postmortem:    [22]   R. D. Galliers and F. F. Land, “Choosing
       Never Leave a Project without it,” IEEE Software,             Appropriate     Information    Systems     Research
       vol. 19, pp. 43-45, 2002.                                     Methodologies,” Communications of the ACM, vol.
[14]   F. D. Davis, “A Technology Acceptance Model for               30, pp. 900-902, 1987.
       Empirically Testing New End-User Information           [23]   T. DeMarco and T. Lister, Peopleware: Productive
       Systems: Theory and Results,,” in MIT Sloan                   Projects and Teams, 2. ed. New York: Dorset House
       School of Management. Cambridge, MA, USA: MIT                 Publishing, 1999.
       Sloan School of Management, 1986.                      [24]   Y. Malhotra and D. F. Galletta, “Extending the
[15]   B. A. Kitchenham, “Evaluating Software                        Technology Acceptance Model to Account for
       Engineering Methods and Tools,” ACM SIGSoft                   Social Influence: Theoretical Bases and Empirical
       Software Engineering Notes, pp. 11-15, 1996.                  Validation,” in Proc. Thirty-Second Annual Hawaii
[16]   E. Metzker and M. Offergeld, “An Interdisciplinary            International Conference on System Sciences, 1999,
       Approach for Successfully Integrating Human-                  pp. pp. 1006.
       Centered Design Methods Into Development               [25]   A. Cockburn, Agile Software Development:
       Processes Practiced by Industrial Software                    Addison Wesley Longman, 2001.
       Development Organizations,” in Engineering for         [26]   N. Juristo and A. M. Moreno, Basics of Software
       Human Computer Interaction: 8th IFIP International            Engineering Experimentation. Dordrecht: Kluwer
       Conference, EHCI 2001(EHCI'    01), Lecture Notes in          Academic Publishers, 2001.
       Computer Science, R. Little and L. Nigay, Eds.         [27]   Standish Group. “CHAOS Chronicles or CHAOS: A
       Toronto, Canada: Springer, 2001, pp. 21-36.                   Recipe For Success”. 1995.
[17]   L. L. Constantine and L. A. D. Lockwood, Software      [28]   Bayer, J. and Melone, N. Adoption of Software
       for Use: A Practical Guide to the Models and                  Engineering Innovations in Organizations. Technical
       Methods of Usage-Centered Design: Addison-                    Report CMU/SEI-89-TR-017, Software Engineering
       Wesley, 1999.                                                 Institute, Carnegie Mellon University.
[18]   D. J. Mayhew, The Usability Engineering Lifecycle:     [29]   Desmarais M.C., Leclair R., Fiset J.Y., Talbi H.
       A Practioner' s Handbook for User Interface Design:           “Cost-Justifying Electronic Performance Support
       Morgan Kaufman Publishers, 1999.                              Systems.” Communications of the ACM, Vol. 40,
[19]   R.     Spencer,    “The    Streamlined    Cognitive           No. 7, July 1997.
       Walkthrough Method: Working Around Social              [30]   Howard R. “Software Process Maturity: Measuring
       Constraints Encountered in a Software Development             Its Impact on Productivity and Quality.” IEEE
       Company,” in Proc. Conference on Human Factors                International Software Metrics Symposium, May
       in Computing Systems (CHI'  00), 2000, pp. 353-359.           1993.