Proceedings of EMOOCs 2019: Work in Progress Papers of the Research, Experience and Business Tracks Quality Assurance Methods Assessing Instructional De- sign in MOOCs that implement Active Learning Peda- gogies: An evaluative case study Valeria Aloizou1, Sara Lorena Villagrá Sobrino2, Alejandra Martínez Monés2, Juan I. Asensio-Pérez2, Sara García Sastre2 1 University of Piraeus, Piraeus, Greece 2 GSIC/EMIC Research Group, Universidad de Valladolid, Valladolid, Spain aloizouv@gmail.com, sarena@pdg.uva.es, amartine@infor.uva.es, juaase@tel.uva.es, saragar@pdg.uva.es Abstract. Since more and more Massive Open Online Courses (MOOCs) ap- pear, constituting an innovative part of online education, they should also un- dergo quality assurance to assess their pedagogical design similarly to other online courses. Recently, new studies have appeared presenting quality assur- ance methods (QA) for assessing the instructional design in MOOCs with peda- gogical innovations, like those implementing active learning pedagogies (e.g. collaboration, gamification), which provide designers and instructors with use- ful strategies in order to design more interactive MOOCs. However, it is not that clear how these efforts are being addressed and up to which point are ap- propriate to assess active learning pedagogies. Therefore, a Systematic Litera- ture Review (SLR) was conducted to identify the most mature existing MOOC QA methods. Afterwards, an evaluative case study was carried out, based on the Evaluand-oriented Responsive Evaluation Model (EREM), to apply the selected methods to a MOOC implementing active learning pedagogies. As the results suggest, the instruments of the selected QA methods need enrichment to assess effectively the instructional design based on active learning pedagogies, by providing specific questions proper for this kind of MOOCs or, by stating the underlying pedagogical model clearly so that the designers could consider be- forehand if it is appropriate or not for their case. The results of the study are a first step to define new, enriched quality assessment methods for MOOCs that apply active learning approaches. Keywords: MOOCs, Quality Assurance Methods, Quality Frameworks, In- structional Design, Active Learning Pedagogies 1 Introduction Millions of people are learning in hundreds of Massive Open Online Courses (MOOCs). MOOC learners are a vast online learning community with diverse moti- vational interests; therefore MOOCs should also undergo quality assurance like other online courses [1]. Recently, some new studies have appeared presenting quality as- surance (QA) methods assessing specific aspects of the instructional design in MOOCs, such us: collaboration, feedback course overview, learning objectives, as- 14 Proceedings of EMOOCs 2019: Work in Progress Papers of the Research, Experience and Business Tracks sessment, instructional materials, learner interaction and engagement, learner support and accessibility [2,3]. However, most existing QA approaches for MOOCs do not cater for the increasing use of active learning pedagogies (e.g. collaboration, gamification), that aim at proposing richer learning experiences that go beyond one- size-fits all approaches [4-6]. This paper addresses this research gap in the QA for MOOCs implementing active learning pedagogies, by exploring the following re- search question (RQ): How can we assess the instructional design quality of a MOOC implementing active learning pedagogies? In order to answer this RQ, we firstly car- ried out a Systematic Literature Review (SLR) to identify and select the most mature existing MOOC QA methods. Secondly, an evaluative case study was carried out, based on the Evaluand-oriented Responsive Evaluation Model (EREM), to apply the selected methods to a MOOC implementing active learning pedagogies and evaluate their instruments. Finally, the findings and the results of the data analysis are report- ed, followed by the main conclusions and recommended lines for future work. 2 Related Work Following the RQ question of the present study, we firstly searched the terrain of QA in MOOCs through an SRL [7] aiming to find out: (i) which QA methods have been proposed so far? (ii) which are their main strengths and weaknesses while assessing the instructional design of a MOOC? Specific electronic databases were selected; IEEE Xplore Digital Library1, Springer Link2, ACM Digital Library3 and Google Scholar4, and particular search strings were used: “MOOCs quality”, “pedagogical quality in MOOCs” and “instructional design QA in MOOCs”, including journal pub- lications, conference proceedings, books and book chapters. The inclusion criteria were publications: (i) written in English, (ii) referring to QA and accreditation, (iii) analyzing quality frameworks and (iv) presenting quality indicators. The exclusion criterion was not referring at all to online learning QA. The 10 initially identified QA methods were assessed to find out which of them better fulfilled the following crite- ria: (i) focus on MOOCs, (ii) inclusion of assessment instruments, (iii) focus on in- structional design, (iv) evidence of testing of the framework, (v) inclusion of the pro- cess/methodology of analysis, and (vi) assessment of active learning pedagogies. 3 QA methods fulfilled most of the criteria and were finally selected in order to proceed with the case study; the Ten-principle framework [8], the OpenupED Quality Label [9] and Quality Matters [10]. 3 Methodology 3.1 Context In order to address the RQ in real contexts, we selected the “Innovative and Collabo- rative Learning with ICT” MOOC (CLAT MOOC) carried out by the University of 1 IEEE Xplore Digital Library, available at http://ieeexplore.ieee.org/Xplore/home.jsp . 2 Springer International Publishing, available at https://link.springer.com/ . 3 ACM Digital Library, available at https://dl.acm.org/ . 4 Google Scholar, available at https://scholar.google.gr/ . 15 Proceedings of EMOOCs 2019: Work in Progress Papers of the Research, Experience and Business Tracks Pompeu Fabra5 and the University of Valladolid6. The course was designed based on the principles of active learning pedagogies, applying innovative elements in terms of collaboration and gamification, such as innovative group formation policies and col- laborative quizzes leading to badge acquisition. CLAT MOOC was a 6-week-course, with 759 total enrollments, delivered by 4 teachers. The course was targeted to inno- vative pre-service and in-service teachers interested in incorporating collaboration with technology into their own teaching practices. 3.2 The case study The assessment tools of the 3 finally selected QA methods were applied in a real case study of the CLAT MOOC. The design of the evaluative case study was based on the Evaluand-oriented Responsive Evaluation Model (EREM), which highly encourages the plurality of data gathering techniques and informants in order to obtain different perspectives about the Evaluand, thus enriching the evaluation process [11]. We as- sumed that this research method was the most appropriate, due to the mixed data gathering techniques and the multiple informants needed in this phase of the study. The Issue (the evaluative tension in EREM terminology) of the case study was de- fined as: To what extent did the selected QA methods help assess the instructional design quality and the active learning pedagogies of the CLAT MOOC? To help illu- minate our research question, we performed an anticipatory data reduction process during the evaluation design (see figure 1), where the Issue was split into more con- crete topics (T). Each topic was explored through a number of Informative Questions (IQ), in order to give answers to the issue of the CLAT MOOC study. Figure 1. Anticipatory data reduction diagram showing the Issue, T and IQ. 5 Universitat Pompeu Fabra Barcelona: https://www.upf.edu/en/ 6 Universidad de Valladolid: http://www.uva.es/ 16 Proceedings of EMOOCs 2019: Work in Progress Papers of the Research, Experience and Business Tracks Multiple data gathering techniques were used (see table 1). Firstly, we analyzed the design of the CLAT MOOC in order to understand its context, which was en- riched with data gathered through informal interviews with the CLAT MOOC design- ers. Secondly, we carried out an evaluation of the quality tools, by conducting one questionnaire and a focus group, addressed to seven (7) research experts in learning design and assessment in MOOCs. Important findings emerged after the content anal- ysis of the transcription and the notes, which are presented in the following section along with the supporting evidence [12]. Table 1. Data sources and labels used to quote them along the study Data N Description of technique Label Sources CLAT 10 Analysis of the CLAT MOOC instructional design to [CLAT] MOOC understand how the active learning strategies were design designed and implemented. Informal 2 Informal interviews with CLAT MOOC designers to [INT] Interviews analyze the instructional design of the CLAT MOOC. Question- 7 Questionnaires provided to research experts, to gather [QUEST] naire their initial opinion regarding the QA tools after their application on the CLAT MOOC. Focus 7 Focus group with research experts to gather their [FOC] group overall opinion about the QA methods. 2.3 Findings The main findings derived from the data analysis in terms of the Issue and its corre- sponding topics to help illuminate the IQ are followingly presented. T1: According to participants, assessment tools could be useful since they: (i) provide guidelines to MOOC designers, (ii) help MOOC designers to identify gaps during the design pro- cess, (iii) could work as a good strategy while re-designing a MOOC, considering the results of the quality assessment (see Table 2, [QUEST]-A). T2: Participants were not satisfied with the questions referring to the collaborative activities and active learning pedagogies, supporting that there is a room for further improvement (see Table 2, [QUEST]-B). T3: The effort which was required while applying the quality tools was assessed according to the time consuming requirements, a factor which is affected either by the comprehensible vocabulary, or by the quality tools layout (see Table 2, [QUEST]-C). T4: Participants stated that a QA method by itself cannot help them assess the objectives that a designer sets (see Table 2, [FOC]-A). Also, it was stated the idea of forming more specific questions with special emphasis in the reflec- tion on pedagogical aspects such as collaboration. In this way the quality of particular instructional design initiatives could be assessed more efficiently (see Table 2, [FOC]-B). T5: Most of the participants considered that QA methods were useful since they could reflect on important elements of their MOOCs. However there is space for improvement (see Table 2, [QUEST]-D). T6: Lastly, it was pointed out that one of the main challenges that MOOC teachers’ face is the assessment of students’ final products due to the massive scale of the courses (see Table 2, [FOC]-C). 17 Proceedings of EMOOCs 2019: Work in Progress Papers of the Research, Experience and Business Tracks Table 2. Selected excerpts of evidence Data Excerpts Source [QUEST] A. The QA tools were useful to identify aspects that should be taken into ac- count when designing a MOOC, and possible weaknesses that have to be ad- dressed. For example, in our case, the aspects related to assessment and the institutional support seems to be the weakest aspect. B. The questions regarding active pedagogies were mainly focused on partici- pation and collaboration, but there are many other strategies that promote active learning and were not evaluated (e.g., inquiry-based learning, problem- solving, role-playing, game-based learning or gamification). C. For me Ten-Principle framework dealt with the instructional design and was easier for me to answer its questions. The other two had some questions that I did not find easy to answer. D. I’d say they helped me to realize that the instructional design had taken into account some “basic” elements of any learning design (e.g., stating objectives, defining assessment criteria, etc.). On the other hand, the assessment of feed- back is not very well developed. Important aspects regarding the assessment that affects the quality design in a MOOC are not informed in this instrument. [FOC] A. I can’t expect from any instrument to say that this objective is correct, be- cause it depends on each MOOC context. B. For example, for the collaborative learning, I would expect questions such as: “Did you choose specific criteria for forming the groups?” C. Also, we couldn’t assess the student’s final products. We couldn’t provide real assessment for the objectives after the activities submission. 3 Conclusions and future work The work presented in this paper has explored the area of QA in MOOCs, and its findings have provided initial insights into how the most significant existing methods can be enriched to assess effectively the instructional design in MOOCs implementing active learning pedagogies. To sum up, there is indeed a need of QA methods in order to: (i) detect weaknesses and elements of the course that need improvement, (ii) as- sess elements that have not been taken into account while designing, (iii) acquire important information while re-designing a MOOC. The QA methods should report clearly the underlying pedagogical model, including clear and simple questions as- sessing as well specific elements of the active learning pedagogies. Thus, designers could end up with more valid and accurate conclusions about the quality of their MOOCs. A line of future work opened by this research is to consider the above men- tioned insights and apply them to the definition of a new, enriched QA method for assessing MOOCs. Acknowledgments This research has been partially funded by the European Regional Development Fund and the National Research Agency of the Spanish Ministry of Science, Innova- tions and Universities under project grants TIN2017-85179-C3-2-R and TIN2014- 18 Proceedings of EMOOCs 2019: Work in Progress Papers of the Research, Experience and Business Tracks 53199-C3-2-R, by the European Regional Development Fund and the Regional Min- istry of Education of Castile and Leon under project grant VA257P18, and by the European Commission under project grant 588438-EPP-1-2017-1-EL- EPPKA2-KA. The authors thank the rest of the GSIC-EMIC research team for their valuable ideas and support. References 1. Gamage, D., Fernando, S., & Perera, I.: Quality of MOOCs: A review of literature on effectiveness and quality aspects. Ubi-Media Computing (UMEDIA). pp. 224-229. IEEE. Colombo, Sri Langa. (2015). doi: 10.1109/UMEDIA.2015.7297459 2. Yousef, M., Chatti, A., Schroeder, U., & Wosnitza, M.: What drives a successful MOOC? An Empirical Examination of Criteria to Assure Design Quality of MOOCs. IEEE 14th International Conference in Advanced Learning Technologies (ICALT). pp. 44-48. Athens, Greece. (2014). doi: 10.1109/ICALT.2014.23 3. Jansen D., Rosewell J. & Kear K.: Quality Frameworks for MOOCs. In: Jemni M., Kinshuk, Khribi M. (eds) Open Education: from OERs to MOOCs. Lecture Notes in Educational Technology. pp. 261-281. Springer. Berlin, Heidelberg. (2017). doi: 10.1007/978-3-662-52925-6_14 4. Bonwell, C.C., Eison, J.A.: Active Learning: Creating Excitement in the Classroom. ASHE-ERIC High. Educ. Rep. No. 1. Washington, DC. (1991). doi:ED340272 5. Khalil, M., Wong, J., de Koning, B., Ebner, M., & Paas, F.: Gamification in MOOCs: A Review of the State of the Art. Global Engineering Education Conference (EDUCON). pp. 1629-1638. IEEE. Tenerife, Spain. (2018). doi: 10.1109/EDUCON.2018.8363430 6. Stracke, C. M.: The Quality of MOOCs: How to improve the design of open educa- tion and online courses for learners? International Conference on Learning and Col- laboration Technologies. pp. 285-293. Springer. Vancouver, Canada (2017). doi: 10.1007/978-3-319-58509-3_23 7. Kitchenham, B., Charters, S.: Guidelines for performing systematic literature reviews in software engineering. Technical report. Ver. 2.3. Durham, UK. EBSE. (2007). 8. Margaryan, A., Bianco, M. & Littlejohn, A.: Instructional quality of Massive Open Online Courses (MOOCs). pp. 77-83. Computers & Education. Glasgow, UK. (2015). doi: 10.1016/j.compedu.2014.08.005 9. Rosewell, J.: Benchmarks for MOOCs: the OpenupEd quality label. International Conference on Enhancement and Innovation in Higher Education. Glasgow, UK. (2015). 10. Lowenthal, R. L., & Hodges, C. B.: In Search of Quality: Using Quality Matters to Analyze the Quality of Massive, Open, Online Courses (MOOCs). The International Review of Research in Open and Distributed Learning. pp. 83-100. (2015). doi: 10.19173/irrodl.v16i5.2348 11. Jorrín-Abellán, I. M., Stake, R. E., & Martínez-Moné, A.: The needlework in evaluat- ing a CSCL system: The Evaluand oriented responsive evaluation model. In Proceedings of the 9th international conference on Computer supported collabora- tive learning. pp. 68-72. International Society of the Learning Sciences. Rhodes, Greece. (2009). 12. Miles, M. B., & Huberman, A. M.: Qualitative data analysis: An expanded source- book. SAGE. New Delhi, India. (1994). 19