=Paper=
{{Paper
|id=Vol-1738/IWTA_2016_paper8
|storemode=property
|title=Improving Personalized Feedback at the Workplace with a Learning Analytics enhanced E-portfolio
|pdfUrl=https://ceur-ws.org/Vol-1738/IWTA_2016_paper8.pdf
|volume=Vol-1738
|authors=Marieke van der Schaaf,Geraldine Clarebout
|dblpUrl=https://dblp.org/rec/conf/ectel/SchaafC16
}}
==Improving Personalized Feedback at the Workplace with a Learning Analytics enhanced E-portfolio==
Improving Personalized Feedback at the Workplace with a Learning Analytics enhanced E-portfolio M. van der Schaaf G. Clarebout Utrecht University, The Netherlands Maastricht University, The Netherlands 3508 TC Utrecht, The Netherlands +31 (0)30 2534944 m.f.vanderschaaf@uu.nl independently executable within a time frame, observable ABSTRACT and measurable in their process and outcome, and, therefore, suitable for entrustment decisions. This is a During workplace based learning, e.g. clinical practice or promising route that is now being explored and during an internship, there is an urgent need for solutions implemented in several countries across the globe (e.g. to restore and to guarantee the quality of feedback for USA, Canada, Australia, Singapore, The Netherlands). (becoming) professionals. In continuing education at the workplace the use of Electronic portfolios (EPs) is a So far the implementation of E-portfolios in workplace- crucial means for acquiring the requisite professional based learning is often ineffective; its quality (in terms of knowledge and skills. Although EPs provide a useful validity and reliability) is generally low and moreover the approach to view each trainee’s progress, often only impact on learning is limited (Van Schaik, Plan, & limited use is made of the rich contextual learning O’Sullivan, 2013). This seems especially the case when assessment data to support responsive adaptation for the E-portfolios are not tailored to show what really more efficient and rewarding training and hence to happened in the workplace (Van der Schaaf, Stokking, & provide personalized feedback. This contribution Verloop, 2008). Part of this failure may be attributed to a advocates that EPs enhanced with a Learning Analytics wish to translate competencies, designed as rather engine, may increase the quality and efficiency of theoretical descriptions of professionals, into items in a workplace-based feedback and assessment. This portfolio for assessment. Furthermore, potential data contribution addresses this by outlining an approach that about trainees’ behaviour in the workplace are often is applied in a European 7th framework project, called underused, because the management of the data is too WATCHME (www.project-watchme.eu). The aim of the complex for the trainees and their supervisors. This paper contribution is to provide insight in underlying rationales addresses this by outlining an iterative development to improve workplace-based feedback and assessment approach that is applied in a European 7th framework and how this is applied in an EP environment that is project, called WATCHME (www.project-watchme.eu). enhanced with Learning Analytics. The project uses an E-portfolio system that is enhanced with a Learning Analytics (LA) engine to provide Keywords personalized (just-in-time) assessment and feedback. LA Learning analytics; workplace-based learning; include the measurement, collection, analysis and competencies; electronic portfolios. reporting of data about learners and their contexts, for purposes of understanding and optimising learning and 1. INTRODUCTION the environments in which it occurs (Clow, 2013; Feedback at the workplace is crucial for trainees to Ferguson, 2012; Siemens & Long, 2011). become professionals. Paralleling the movement towards The design approach for the LA engine that drives the E- alternative assessments of students (Boud, 1990; portfolio is of a cyclical nature based on ongoing Birenbaum 1996), (becoming) professionals are refinement and improvement of the engine during increasingly assessed using competence-based successive phases of collection, analysis and visualising assessment instruments, such as portfolios. A portfolio information (Baker & Yacef, 2008; Elias, 2011). Though contains selected evidence of trainees’ learning LA are driven by a computerised processing of large processes, their performances and products in various amounts of data, the analytical process is a ”single contexts, accompanied by supervisors’ comments and amalgam of human and machine processing which is reflections (Wolf & Dietz, 1998). Depending on its instantiated through an interface that both drives and is content and mode of presentation an electronic portfolio driven by the whole system, human and machine” (Dron (E-portfolio) can do justice to the fact that professional & Anderson, 2009, p. 369). Student Models will be used practice is complex and context dependent. as a means of analysis, the results of which will lead to In this paper we use Entrustable Professional Activities two types of feedback: Just-in Time feedback messages (EPAs) to describe units of professional practice that and visualization of both individual and aggregated data. underlie workplace-based feedback and assessment In order to provide meaningful just-in-time information, (Gilhooly, Schumacher, West & Jones, 2014; Jones, the Student Model should represent the actual internal Rosenberg, Gilhooly, & Carraccio, 2011; Ten Cate, state of each trainee as well as their actual learning 2013). EPAs are tasks or responsibilities entrusted to be context. For this, it must be able to interpret the contents executed by an unsupervised learner once sufficient of the E-portfolio. The Student Model should also contain specific competence has been obtained. EPAs are enough pedagogical knowledge in order to be able to ELECTRONIC PORTFOLIOS FOR WORKPLACE-BASED FEEDBACK 2 translate the internal state and context into meaningful perform at several EPAs at the workplace and on a messages and information for visualization. Key in second level reveals their performance on the underlying enhancing E-portfolios with LA is that data about competencies. The personalized feedback aims to give trainees’ workplace performances are linked to trainees insight into their obtained score compared with assessment and feedback scores. This requires the the expected norm (they can infer at what entrustment alignment of a statistical model with a substantive theory, level they are), it provides them the chance to evaluate operationalized in EPA descriptions, regarding expertise and monitor the own process (trainees need to reflect development in the profession. To this end, an iterative upon their performance) and the exemplar performance development approach, using various cycles will be (example feedback) gives suggestion upon how to close applied. the gap between the expected norm and the actual performance. Hence, the feedback is based upon the three The aim of this contribution is to develop a design for principles of effective feedback and uses exemplar personalized feedback in a LA-driven E-portfolio. The performance (Sadler, 1989; 2010). central question is: How can a LA-enhanced E-portfolio improve feedback at the workplace to enhance 3. Student Model (becoming) professionals’ development? Decisions on entrustability (or proficiency) levels for EPAs are made on the basis of a set of workplace-based 2. Personalized Feedback assessments, not using strict addition of scores but using High quality feedback is essential to stimulate rich, partly narrative, information. This means that a crisp (becoming) professionals’ EPA development. Feedback rule-based approach is not feasible whereas a can be conceptualised as information provided by an probabilistic approach is able to deal with the agent regarding aspects of one’s performance or uncertainties in this type of decision making. The understanding. For feedback to be effective certain underlying Student Model needs to be able to advice on conditions must hold; the feedback must be given timely (at least): and adequately, it needs to be of high quality, and learners should be able to act upon the feedback (Gibbs & 1. Prediction of entrustability: What is, at this moment, Simpson, 2004). Furthermore, there is a large body of probably the current level of research to show that good feedback leads to achieve entrustability/proficiency for a trainee in a given aimed performances (Nicol & Macfarlane-Dick, 2006). EPA? This can be expressed as a probability At least three conditions should be fulfilled for feedback distribution over the levels x for that EPA given the to be effective: 1) it gives insight into obtained current evidence: performances compared to an expected norm, 2) it gives 𝑃(𝑙𝑒𝑣𝑒𝑙 𝑥 𝑓𝑜𝑟 𝐸𝑃𝐴 | 𝑐𝑢𝑟𝑟𝑒𝑛𝑡 𝑒𝑣𝑖𝑑𝑒𝑛𝑐𝑒 𝑖𝑛 𝑝𝑜𝑟𝑡𝑓𝑜𝑙𝑖𝑜) the ability to evaluate and monitor the own process and 3) If feasible, a Value-of-Information analysis can be it gives suggestions to fill the gap between the expected performed to identify the unknown variables that norm and the actual performance (Sadler, 1989; 2010). would provide the most information to answer. Hence, helpful feedback states what aimed performances 2. Selection of feedback: What is the best feedback to are and how current performance is related to the select for a given trainee at a given moment? performances aimed at. Further, it provides action points 3. Selection of topic of interest: What EPA, task or on how to close the gap between current and aimed competency is at the moment the most of interest for performance. Furthermore, effective feedback enhances trainee/supervisor? learning when it provides answers to the following question: Where am I going? How am I going? and Where to next? (Hattie & Timperley, 2007). It is thus important that trainees get acquainted with the goals and ‘criteria’ of an EPA, infer how they performed and know how to enhance their performance. Trainees can only achieve development goals when they understand those goals and can assess their progress (Sadler, 1989). One approach that is particularly powerful in clarifying goals and standards has been to provide trainees with rubrics (Dekker-Groen, Van der Schaaf & Stokking, 2012). Rubrics can be effective because they make explicit what is required of trainees’ performance, they define a valid standard against which trainees can compare their work and hence, may enhance trainees’ further learning. This contribution focuses on providing trainees personalized feedback on the process of becoming a professional. The feedback module is based on EPAs that go with rubrics that describe entrustability or proficiency levels. It consists of a personalized feedback module (JIT) and a visualization module (VIZ). This JIT and visualization uses Student Models to depict how trainees’ ELECTRONIC PORTFOLIOS FOR WORKPLACE-BASED FEEDBACK 3 that describes the educational context is specific for each trainee and needs to be re-constructed frequently, since 4. Designing LA-enhanced Electronic the actual educational context changes continuously. Given the high levels of uncertainty in the educational Portfolios domain, probabilistic approaches are appropriate and The design of a LA-enhanced E-portfolio in our project graphic models such as Bayesian networks support the demanded interdepended phases in which the involved modular structure most appropriately. Before SMs can be educational and technical partners have to answer developed, different questions have to be answered specific questions. amongst the users, e.g.: When do users require feedback? Phase 1. Development of EPAs and assessment How do users perceive feedback? What timing of instruments. Phase 1 started the development cycle by feedback is useful? defining the competencies needed and types of evidence Phase 3. Development of initial Personalized (e.g. products and performances) that should go in the E- Feedback Module. This phase addressed the portfolios for valid workplace-based assessment. Users development of initial Personalized feedback module that (experts and trainees) are consulted to generate markers produces, on the basis of information retrieved from for progress within the professional domain and Student Models, feedback to trainee and supervisors. consensus will be sought to arrive at generalizable Also visualization modules (VIZ) are developed that will, weighted markers that will be suitable to translate to on the basis of information retrieved from Student Learner Analytics input, i.e. the “Student Models” in Models and portfolio data, produce informative graphical phase 3. Main questions to be answered are: what representations of aggregated and individual data, see competencies need to be assessed and what types of Figure 2. The detailed designs of JIT and VIZ demand evidence (e.g. product, performance, processes) should input from the users on questions like: What kind of go in the E-portfolios? In previous studies, in which we feedback do they prefer, with what graphically display? used a Delphi technique (Linstone & Turoff 1975), What are the time constraints for giving and receiving stakeholders successfully developed EPAs for the feedback? What kinds of devices are available when professional fields of medical education, veterinary assessment is performed and received? The personalized education and teacher education. See Figure 1 for an feedback module will be accessible from the E-portfolio, example of teacher education. representing the output of the underlying SMs. The SM is Phase 2. Development of Student Models. Phase 2 took a back-end service in itself and is not available for user the output of phase 1 and technical considerations, such interaction in the display, but the JIT and VIZ that are as scalability, into account. Educational mining tools and driven by SM are. See Figures 2a-2c. These figures show techniques are selected that will be deployed to learn, a possible example of personalized feedback and EPAs update and store the Student Models. Student Models attained. The personalized feedback is dynamic and (SMs) are statistical models that predict trainees’ continually receives input from new incoming portfolio progress based on existing data. They translate the data. The final display knows several layers providing portfolio and assessment data into the progress state of extra detailed information when one clicks on a certain the trainee. As a consequence SMs will predict the actual graph, message etc. in the display. state of performance of each trainee within their actual workplace based learning context. The part of the SM EPA 1. Sets learning goals for the whole curriculum and specific lessons Assessment and The teacher does/does not formulate (self formulated) learning goals in connection with specific evaluation criteria subject content The teacher does/does not make use of SMART (specific, measurable, acceptable, realistic and time related) formulated learning goals. The teacher does/does not take into consideration the starting situation of students when formulating learning goals. Proficiency levels The teacher takes over the learning goals or course material from others. He/she incidentally considers the starting situation of the students and the connection with specific subject content. The teacher does not check if the learning goals are SMART formulated. (starting) The teacher regularly checks if the learning goals of others or the course material connect to specific subject content and the starting situation of the students. The teacher checks if the set learning goals are SMART formulated. (sufficient) The teacher formulates his/her own learning goals, which usually connect to the specific subject content and the starting situation of the students. These learning goals are partially SMART formulated. (good) The teacher formulates his/her own coherent learning goals, which connect to the specific subject content and the investigated starting situation of the students. The learning goals are SMART formulated. (Excellent) Assessment forms Lesson plans/series of lessons and student placement evaluation form. Assessor Institute and internship supervisor. Figure 1. Rubrics in Teacher Education Figures 2a-2c. Personalized feedback at Entrustment level 5. Rationale of Personalized Feedback standards. This is visualized in the overviews with scores The personalized feedback module that we developed in on EPAs and competencies (see Figures 3 and 4). the project is inspired by Nicol and MacFarlane-Dick’s 2. Facilitates the development of self-assessment seven principles of good feedback practice (2006) that (reflection) in learning. Our design allows for close facilitate self-regulation. These principles were translated monitoring of trainees’ progress by visualizing trainees’ in the design as follows. Good feedback: performance on the EPAs by means of graphs and figures 1. Helps clarify what good performance is . For as well as narrative feedback. In this way it provides an professional development at the workplace the learning overview of students’ strengths and points for further goals should be crystal clear in order to stimulate learning development, which can be used for self-assessment and and above that should stimulate (learn) trainees to clarify peer assessment and discussion about trainees’ portfolio. own goals (Sadler, 1989). It is well known that often Further, compiling the portfolio (selecting materials as mismatches occur between supervisors’ and trainees’ input for the portfolio) already demands trainees’ interpretation of assessment criteria and standards, reflection. especially when it comes down to complex tasks at the workplace that can be tacit and culture related. An approach that we provided is the development of EPAs connected in rubrics (see Figure 1). Rubrics have proven to be very helpful in clarifying goals and standards and stimulating trainees in goal clarification and goal setting, for instance by involving trainees in the assessment and stimulating discussion and reflection about criteria and ELECTRONIC PORTFOLIOS FOR WORKPLACE-BASED FEEDBACK 5 Figure 3. Spider chart view of scores on the EPAs Figure 4. Spider chart view of scores on competencies Figure 5. Timeline view of trainees’ performance on EPAs (called tasks in this example). ELECTRONIC PORTFOLIOS FOR WORKPLACE-BASED FEEDBACK 6 3. Delivers high quality information to trainees about Example of Aggregated Feedback their learning. Trainees need detailed information of Feedback Type message (Level 1) high level to monitor and correct their own performance Improvement There is room for improvement for and to take action to improve. In the preliminary this EPA. Please click on the personalized feedback module this is enhanced by: (a) message to see how you can improve linking the feedback to predefined EPAs that includes your performance. criteria and standards; (b) ensuring timely feedback by Positive You have recently received good means of underlying SMs that feed into the system; (c) scores for this EPA. Please click giving trainees advise on their learning and showing here to see how you can improve (prioritizing) needs for improvement; (d) regulating the more. amount of feedback by giving trainees the option to click Trend You currently have a trend of further if they want more detailed information; (e) decreasing scores for this EPA. allowing to upload information in the portfolio system Supervisor Your supervisor added few anytime anywhere, which makes the feedback system up improvement comments on this EPA. to date. See Figure 5 for examples of types of feedback. Cohort Compared to your cohort, you 4. Encourages teacher and peer dialogue around received better scores than your learning. The system allows for supervisor and peer peers on this EPA. dialogues about progress and possible improvement. Gaps You have less assessments than your Such dialogues are important to make sure that trainees peers on this EPA. understand the feedback, can value and verify it and know how to act on it (Van der Schaaf et al., 2008). The Feedback Type Some examples of Detailed E-portfolio environment allows for interaction between Feedback message (Level 2) supervisors, trainees and peers and has the possibility that Improvement You are level 2 on your Physical several stakeholders upload documents, so that for Examination Competency. To instance peer feedback can be used as ‘evidence’ for a achieve the next level your trainee’s performance. examination and research should be reasonably complete and technically 5. Encourages positive motivational beliefs and self- esteem. Positive motivational beliefs and self-esteem are adequate. Overview of the prerequisite for learning and improved performance. It is examination and consistency are known that both benefit most when trainees receive many reasonably developed. low-stakes assessment tasks, with immediate feedback Trend You were level 3 on your Physical for improvement (if needed), rather than receiving few Examination Competency and you high-stakes summative assessment tasks. The E-portfolio dropped on level 2 during your last allows the trainee to select and rewrite own pieces of assessment. To achieve the next work/documents (drafts and resubmissions) and level your examination and research formative feedback in de long run. The SM instantly should be reasonably complete and updates when new information comes in. technically adequate. Overview of 6. Provides opportunities to close the gap between the examination and consistency are current and desired performance . Feedback in the EP reasonably developed. should support trainees to take the next steps to improve Supervisor "You are performing well, but you their performance. This demands engagement for further can take more notes during the improvement and can be stimulated by providing examination process." (13/05/2015) feedback on work in progress, provide feedback in several stages in which feedback (Gibbs, 2004). The E- Figure 6. Examples of detailed feedback messages for portfolio allows this. each feedback type 7. Provides information to teachers that can be used to help shape the teaching. Not only trainees need to be 6. Discussion informed about their progress and options for The aim of this contribution was to elucidate how improvements, this also counts for the supervisors. They personalized feedback based upon Learning Analytics need to be informed with detailed and quality information could be used in an E-portfolio environment. The E- about their trainees in order to guide them at the portfolio offers learners (students, trainees, professionals) workplace. This especially counts for professional and their supervisors an environment to monitor and education in which trainees have many supervisors for provide evidence of their learning and competency several internships. These supervisors often do not know development. The progress of the user can be closely what feedback a trainee received from previous monitored by choosing from amongst several display supervisors and how trainees’ longitudinal progress looks modes, such as radar, line and bar charts, which are like. The preliminary personalized feedback design feeds automatically generated by the system. The scores (on the into this by a specific portfolio entry for supervisors with different competencies) used for these visualizations are long term information about the trainee and the digital abstracted form the assessment tools inserted in the option for trainees to ask for supervisor feedback. portfolio. Other overviews are also displayed, for example numerical overviews of the total inserted forms and an overview of the progress, based on all activities, forms and procedures linked to it. The developed LA- tools will be open source. ELECTRONIC PORTFOLIOS FOR WORKPLACE-BASED FEEDBACK 7 7. ACKNOWLEDGMENTS [8] Ferguson, R. (2012). Learning analytics: drivers, This study was conducted within the framework of developments and challenges. International Journal “Workplace-Based e-Assessment Technology for of Technology Enhanced Learning, 4(5), 304-317. competency-Based Higher Multi-Professional Education” [9] Gibbs, G., & Simpson, C. (2004). Conditions under (WATCHME) project supported by the European which assessment supports trainees’ learning. Commission 7th Framework Programme (grant Learning and Teaching in Higher Education, 1, 3-31. agreement No. 619349). [10] Hattie, J., & Timperley, H. (2007). The power of 8. References feedback. Review of educational research, 77(1), 81- 112. [1] Baker, R. S. J. D., & Yacef, K. (2009). The state of educational data mining in 2009: A review and [11] Nicol, D. J., & Macfarlane‐ Dick, D. (2006). future visions. Journal of Educational Data Mining, Formative assessment and self‐ regulated learning: a 1(1), 3-17. model and seven principles of good feedback practice. Studies in Higher Education, 31, 199-218. [2] Birenbaum, M. (1994). Toward adaptive assessment – the students’ angle. Studies in Educational [12] Sadler, D. R. (1989). Formative assessment and the Evaluation, 20, 239-255. design of instructional systems. Instructional Science, 18, 119-144. [3] Boud, D. (1990). Assessment and the promotion of academic values. Studies in Higher Education, 15, [13] Sadler, D. R. (2010). Beyond feedback: Developing 101-111. trainee capability in complex appraisal. Assessment & Evaluation in Higher Education, 35, 535-550. [4] Clow, D. (2013). An overview of learning analytics. Teaching in Higher Education, 18(6), 683 - 695. [14] Siemens, G., & Long, P. (2011). Penetrating the fog: doi:10.1080/13562517.2013.827653 Analytics in learning and education. Educause Review, 46(5), 30-32. [5] Dekker-Groen, A., Van der Schaaf, M., & Stokking, K. (2012). Performance standards for teachers [15] Van Schaik, S., Plant, J., & O’Sullivan (2013). supporting nursing students’ reflection skills Promoting self-directed learning through portfolios development. Journal of Nursing Education and in undergraduate medical education: The mentors’ Practice, 2, 1, 9-19. doi: 10.5430/jnep.v2n1p9. perspective. Medical Teacher, 35, 139-144. doi: 10.3109/0142159x.2012.733832. [6] Dron, J., & Anderson, T. (2009). How the crowd can teach. In S. Hatzipanagos & S. Warburton (Eds.), [16] Van der Schaaf, M., Stokking, K., & Verloop, N. Handbook of research on social software and (2008). Developing and validating a design for developing community ontologies (pp. 1–17). teacher portfolio assessment. Assessment & Hershey, PA: IGI Global Information Science. Evaluation in Higher Education, 33(3), 245-262. Retrieved from www.igi- doi: 10.1080/02602930701292522. global.com/downloads/excerpts/33011.pdf. [17] Wolf, K., & Dietz, M. (1998). Teaching portfolios: [7] Elias, T. (2011). Learning analytics: definitions, purposes and possibilities. Teacher Education processes and potential. Creative Commons Quarterly, 25, 9-22. Attribution 3.0.